Comment author: jbay 14 January 2016 07:22:26AM 0 points [-]

You're welcome! And I'm sorry if I went a little overboard. I didn't mean it to sound confrontational.

Comment author: isionous 14 January 2016 02:48:36PM 0 points [-]

sorry if I went a little overboard. I didn't mean it to sound confrontational.

You didn't. I appreciated your response. Gave me a lot to think about.

I still think there is some value to my strategy, especially if you don't want to (or it would be unfeasible) to give full probability distribution for related events (ex: all the possible outcomes of an election).

Comment author: jbay 13 January 2016 07:56:06AM *  1 point [-]

The calibration you get, by the way, will be better represented by the fact that if you assigned 50% to the candidate that lost, then you'll necessarily have assigned a very low probability to the candidate that won, and that will be the penalty that will tell you your calibration is wrong.

The problem is the definition of more specific. How do you define specific? The only consistent definition I can think of is that a proposition A is more specific than B if the prior probability of A is smaller than that of B. Do you have a way to consistently tell whether one phrasing of a proposition is more or less specific than another?

By that definition, if you have 10 candidates and no information to distinguish them, then the prior for any candidate to win is 10%. Then you can say "A: Candidate X will win" is more specific than "~A: Candidate X will not win", because P(A) = 10% and P(~A) = 90%.

Since the proposition "A with probability P" is the exact same claim as the proposition "~A with probability 1-P"; since they are the same proposition, there is no consistent definition of "specific" that will let one phrasing be more specific than the other when P = 50%.

"Candidate X will win the election" is only more specific than "Candidate X will not win the election" if you think that it's more likely that Candidate X will not win.

For example, by your standard, which of these claims feels more specific to you?

A: Trump will win the 2016 Republican nomination

B: One of either Scott Alexander or Eliezer Yudkowsky will win the 2016 Republican nomination

If you agree that "more specific" means "less probable", then B is a more specific claim than A, even though there are twice as many people to choose from in B.

Which of these phrasings is more specific?

C: The winner of the 2016 Republican nomination will be a current member of the Republican party (membership: 30.1 million)

~C: The winner of the 2016 Republican nomination will not be a current member of the Republican party (non-membership: 7.1 billion, or 289 million if you only count Americans).

The phrasing "C" certainly specifies a smaller number of people, but I think most people would agree that ~C is much less probable, since all of the top-polling candidates are party members. Which phrasing is more specific by your standard?

If you have 10 candidates, it might seem more specific to phrase a proposition as "Candidate X will win the election with probability 50%" than "Candidate X will not win the election with probability 50%". That intuition comes from the fact that an uninformed prior assigns them all 10% probability, so a claim that any individual one will win feels more specific in some way. But actually the specificity comes from the fact that if you claim 50% probability for one candidate when the uninformed prior was 10%, you must have access to some information about the candidates that allows you to be so confident. This will be properly captured by the log scoring rule; if you really do have such information, then you'll get a better score by claiming 50% probability for the one most likely to win rather than 10% for each.

Ultimately, the way you get information about your calibration is by seeing how well your full probability distribution about the odds of each candidate performs against reality. One will win, nine will lose, and the larger the probability mass you put on the winner, the better you do. Calibration is about seeing how well your beliefs score against reality; if your score depends on which of two logically equivalent phrasings you choose to express the same beliefs, there is some fundamental inconsistency in your scoring rule.

Comment author: isionous 13 January 2016 04:50:45PM 0 points [-]

Thank you for your response.

Comment author: jbay 12 January 2016 05:46:49PM 0 points [-]

Yes, but only because I don't agree that there was any useful information that could have been obtained in the first place.

Comment author: isionous 12 January 2016 06:55:36PM 0 points [-]

Could you comment about how my strategy outlined above would not give useful information?

Comment author: jbay 12 January 2016 07:52:38AM *  0 points [-]

"Candidate X will win the election with 50% probability" also implies the proposition "Candidate X will not win the election with 50% probability". If you propose one, you are automatically proposing both, and one will inevitably turn out true and the other false.

If you want to represent your full probability distribution over 10 candidates, you can still represent it as binary predictions. It will look something like this:

Candidate 1 will win the election: 50% probability

Candidate 2 will win the election: 10% probability

Candidate 3 will win the election: 10% probability

Candidate 4 will win the election: 10% probability

Candidate 5 will win the election: 10% probability

Candidate 6 will win the election: 2% probability

...

Candidate 1 will not win the election: 50% probability

Candidate 2 will not win the election: 90% probability

...

The method described in my post handles this situation perfectly well. All of your 50% predictions will (necessarily) come true 50% of the time, but you rack up a good calibration score if you do well on the rest of the predictions.

Comment author: isionous 12 January 2016 02:52:49PM 1 point [-]

The method described in my post handles this situation perfectly well. All of your 50% predictions will (necessarily) come true 50% of the time, but you rack up a good calibration score if you do well on the rest of the predictions.

Seems like you're giving up trying to get useful information about yourself from the 50% confidence predictions. Do you agree?

Comment author: isionous 11 January 2016 11:19:31PM *  2 points [-]

For the problem of how to phrase the 50% confidence predictions, it might be useful to use the more specific proposition (or one that has the smaller uninformative prior probability). For instance, if you have a race between 10 candidates, and you think candidate X has a 50% chance to win, you should phrase your prediction as "Candidate X will win the election" rather than "Candidate X will not win the election".

If you consistently phrase your 50% confidence predictions this way, then your prediction results tell you something useful. If your 50% confidence predictions come true only 10% of the time, then maybe your problem is overconfidence in your reasons to deviate from uninformed priors or overconfidence in concocting reasons to believe a certain hypothesis in the vast hypothesis space.

edit: The biggest weakness of this approach is what do you do when you're choosing between something like "this coin flip will come up heads" and "this coin flip will come up tails"? Or when it is unclear what an uninformed prior would even be.

Comment author: James_Miller 30 July 2015 11:38:53PM *  5 points [-]

In my imagination, there's a terrible accident that leaves someone other than the helmet-wearer paralyzed or dead, and investigators are surprised to see that one driver was wearing... a helmet?? It's almost like he knew he was going to get into an accident -- perhaps even intended to. Certainly, that's what people would think reading the articles about it. Perhaps a jury would, as well.

If this happens to you say you did it because of my blog post. If you pay transportation costs I will even testify for you at a trial.

anti-lock breaks are said to increase risky driving behavior, after all

Economists call this the "Peltzman effect" and it seems robust. It does reduce the social benefits of driving helmets. One economist took the implications of this effect to their logical conclusion and suggested that steering wheels should have a spike pointed at their drivers.

Comment author: isionous 08 August 2015 08:16:54AM 0 points [-]

What do you think of watching car crash compilations to counteract the Peltzman effect?

Comment author: btrettel 30 July 2015 11:58:40PM *  10 points [-]

Good suggestion. I'm a cyclist and sometimes I walk around wearing my bike helmet. Some people give me funny looks, but I don't care. To regularly wear helmets, you have to recognize that avoiding TBI is more important than looking slightly silly or messing up your hair. Unfortunately most people don't get this. I think if helmets become more popular people will find that they don't actually mind the look, as you suggested in another comment.

Risk compensation is one potential problem with wearing a helmet. An example was mentioned in the original post: drivers passing cyclists who wear helmets closer. Also concerning is that cyclists ride more dangerously if they wear helmets (here's an article which I haven't read in full, but it discusses this issue). I imagine seat belts have a similar effect, as might drivers wearing helmets. Based on this idea backwards, I've read a proposal to add a spike to steering wheels to reduce dangerous driving.

Also worth noting is that avoiding long commutes and drives improves safety. Risk (probably) increases monotonically with duration.

One last suggestion would be to drive with a dashcam to make yourself slightly safer. I ride my bike with a very obvious helmet cam and I have observed that some drivers seem to drive safer around me when they notice it. I'm recording them, so if they make a bad move they'll be much worse off in court. I also think the camera makes me ride safer for the same reasons.

Comment author: isionous 08 August 2015 08:00:11AM 3 points [-]

Risk compensation is one potential problem with wearing a helmet... I imagine seat belts have a similar effect, as might drivers wearing helmets. Based on this idea backwards, I've read a proposal to add a spike to steering wheels to reduce dangerous driving.

I recommend watching car crash compilations. They definitely help you feel the risk of driving on an emotional level. I feel I've learned some things that improve my driving safety, but probably the biggest safety gain is I just tend to drive more cautiously due to car collisions being very present in my mind.

Comment author: Lumifer 18 July 2015 03:53:11AM *  1 point [-]

We would run into the same problem for any description of a quale/sensation.

Only if you decide you're defining a sensation and not some physical phenomenon.

The highly contextual nature of color perception is ubiquitous. Human color processing is always making contextual adjustments from scene illumination.

Yes, I understand that very well. But all that tells you is that different definitions will diverge in many cases.

Also, don't you mean objective?

"Subjective" was probably the wrong word. I distinguish:

  • A physical approach which defines color through spectral power distributions

  • A human objective approach which defines color via the tristimulus model (the CIE color space, etc.)

  • A human subjective approach which defines color as a particular perception

The human subjective approach has -- as you have pointed out -- all the issues associated with talking about subjective sensations, that is, they are essentially unobservable and it's very hard to get a good handle on them. That, to me, makes defining color through qualia a definition that isn't useful all that often.

Comment author: isionous 22 July 2015 04:20:14AM 1 point [-]

Only if you decide you're defining a sensation and not some physical phenomenon...That, to me, makes defining color through qualia a definition that isn't useful all that often.

That's the definition used in the overwhelming majority of cases. Careful, technical texts often make it clear that color is a sensation. Even Isaac Newton stressed that "the rays [of light] are not colored".

Even wikipedia goes with the sensation definition of color: "Color...is the visual perceptual property corresponding in humans to the categories called red, blue, yellow, etc...The color of an object depends on both the physics of the object in its environment and the characteristics of the perceiving eye and brain."

In everyday use, when a person says things like "hand me the blue towel", that person usually does not care, know, or even think about reflectance profiles and spectral power distributions. Usually all that person cares about is that the towel "looks blue" to him and the person he's talking to. He'll say "that towel is blue" just like he'll say "that chocolate is bitter".

It's very useful to have definitions that depend on human sensations. You and I are both humans, and we often have conversations with other humans.

Comment author: Stuart_Armstrong 17 July 2015 10:30:25AM 0 points [-]

Cheers! I used blue and green because the grue and bleen example is a standard philosophical one, and it's not hard to make pretty rigorous (as Lumifer suggested, with the radiation frequency, and some outside conditions added to it). The pretty rigorous definition is only a partial match for the subjective blue and green, as you pointed out, and I'll try and make that clearer in any subsequent write-up.

Comment author: isionous 17 July 2015 05:23:35PM 1 point [-]

it's not hard to make pretty rigorous (as Lumifer suggested, with the radiation frequency, and some outside conditions added to it).

Taking "outside conditions" into account to produce an objective definition of color that does a good job of corresponding to human color sensations is actually extremely complex and a very difficult task. Human color sensations are the result of extremely complex and highly contextual processing. I have studied color vision a great deal, and it is very, very common for people to underestimate the complexity and contextual nature of color perception.

Also, you're already conceding that color is not a property of a single object, which would make color a poor example of a property of an object.

Anyway, I'll take your response as a sign that you are comfortable with the problematic nature of your example, and the more pressing concern is playing nice with philosophical tradition/convention. So, I consider the issue closed.

Comment author: Lumifer 17 July 2015 02:35:13PM *  1 point [-]

Thanks :-)

that is not sufficient nor necessary to ensure that the object would typically generate a red color sensation in humans

You are now dealing in circular logic. If you criterion for "red" is a "color sensation in humans", you have already defined red. That's it, we're done. My point is that you do not have to define "red" in terms of human qualia.

Definitions should be judged by their usefulness. Sometimes "a human would call that red" is the right defintion, sometimes "this light peaks at ~700nm" is the right definition. For example, if the sensor in your telescope captured a few photons from a dim distant star, you might call them "red" even if the human visual system will be unable to process these photons (or associate with them the qualia of "red").

it's not too hard to make color an objective creature

It is easy to make color an objective creature -- just define it a particular mixture of wavelengths of visible light. To produce a workable definition of subjective color is much harder -- this is a practical matter in photography and graphics and whole books are written on the subject.

Comment author: isionous 17 July 2015 05:13:09PM *  0 points [-]

You are now dealing in circular logic. If you criterion for "red" is a "color sensation in humans", you have already defined red. That's it, we're done.

We would run into the same problem for any description of a quale/sensation. For example, we would describe/identify nausea, bitterness, and redness in similar ways - it's hard to directly describe sensations, so we often indirectly identify sensations by pointing to conditions that lead to humans experiencing the sensation, or pointing to how the sensation relates to other sensations. That's the unfortunate business of qualia/sensations.

So, the criterion isn't circular, it's just unsatisfying in that we basically end up saying, "in situation X, you will probably feel the sensation I'm talking about, and we've labeled that sensation 'red'".

Definitions should be judged by their usefulness.

Agree 100%. Sometimes you can pretend that color is a simple objective creature and it turn out okay, just like we can use Newtonian physics when relativistic effects are small enough for our purpose.

just define it [color] a particular mixture of wavelengths of visible light

As I said before, you CAN come up with a simple objective definition of color, but it will do a "very poor job of corresponding to human color sensations".

A particular spectral power distribution can lead to many different color sensation depending on visual context, and even the expectation and memory of the human experiencing the color sensation. This fact makes that sort of definition unfit for a lot of purposes.

Look at these two scenes. The left-hand scene contains "blue" pixels that use the same RGB value (0x6e6f73) as the right-hand scene's "yellow" pixels. And in the context of the legend at top, that RGB value produces a third color sensation: gray.

So, any definition of color that only depends on spectral power distributions will not correspond very well with actual human color sensations. The linked picture demonstrates that "the light mixture your monitor produces for an RGB value of 0x6e6f73" is nowhere near enough information to predict what color sensation a human will experience from viewing pixels with that RGB value, even within the limited range of conditions of looking at something on a monitor.

Also, the two-scene picture is not an unusual case. The highly contextual nature of color perception is ubiquitous. Human color processing is always making contextual adjustments from scene illumination. The picture of a fruit basket in this section does a good job of showing how contextual adjustments are the norm. The overwhelming importance of context in color perception massively shrinks the situations where simple objective definitions of color are useful. Treating color as a simple objective creature gets you into trouble fairly quickly.

To produce a workable definition of subjective color is much harder -- this is a practical matter in photography and graphics and whole books are written on the subject.

Yep, which is why I urged the author of the post to choose something other than color as an example of a simple/natural category.

Also, don't you mean objective? The color model work you're talking about is an effort to come up with objective mathematical models that exist outside of human minds (thus considered objective) that give outputs that correspond to sensations that exist inside human minds (those sensations being subjective). I don't want us to get hung up on what "objective" and "subjective" means, but if this conversation continues much more, it might be good to spend a bit of time making sure we understand each other when we use those words.

View more: Next