How big a deal is this? What, if anything, does it signal about when we get smarter than human AI?
It shows that Monte-Carlo tree search meshes remarkably well with neural-network-driven evaluation ("value networks") and decision pruning/policy selection ("policy networks"). This means that if you have a planning task to which MCTS can be usefully applied, and sufficient data to train networks for state-evaluation and policy selection, and substantial computation power (a distributed cluster, in AlphaGo's case), you can significantly improve performance on your task (from "strong amateur" to "human champion" level). It's not an AGI-complete result however, any more than Deep-Blue or TD-gammon were AGI-complete.
The "training data" factor is a biggie; we lack this kind of data entirely for things like automated theorem proving, which would otherwise be quite amenable to this 'planning search + complex learned heuristics' approach. In particular, writing provably-correct computer code is a minor variation on automated theorem proving. (Neural networks can already write incorrect code, but this is not good enough if you want a provably Friendly AGI.)
The interesting thing about that RNN that you linked that writes code, is that it shouldn't work at all. It was just given text files of code and told to predict the next character. It wasn't taught how to program, it never got to see an interpreter, it doesn't know any English yet has to work with English variable names, and it only has a few hundred neurons to represent its entire knowledge state.
The fact that it is even able to produce legible code is amazing, and suggests that we might not be that far of from NNs that can write actually usable code. Still some ways away, but not multiple decades.
DeepMind's go AI, called AlphaGo, has beaten the European champion with a score of 5-0. A match against top ranked human, Lee Se-dol, is scheduled for March.