I see two major challenges (one of which leans heavily on progress in linguistics). I can see there being mathematical theory to guide candidate model decompositions (Challenge 1), but I imagine that linking up a potential model decomposition to a theory of 'semantic interpretability' (Challenge 2) is equally hard, if not harder.
Any ideas on how you plan to address Challenge 2? Maybe the most robust approach would involve active learning of the pseudocode, where a human guides the algorithm in its decomposition and labeling of each abstract computation.
The following is a hypothesis regarding the purpose of counterfactual reasoning (particularly in humans). It builds on Judea Pearl's three-rung Ladder of Causation (see below).
One important takeaway from this hypothesis is that counterfactuals really only make sense in the context of computationally bounded agents.
Counterfactuals provide initializations for use in MCMC sampling.
Preliminary Definitions
Association (model-free):
Pr(Y=y∣X=x)
Intervention/Hypothetical (model-based):
Pr(Y=y∣do(X=x))
Counterfactual (model-based):
Pr(Y=y∣do(X=x),Y=y′)
In the counterfactual, we have already observed an outcome y′ but wish to reason about the probability of observing another outcome y (possibly the same as y′) under do(X=x).
Note: Below, I use the terms "model" and "causal network" interchangeably. Also, an "experience" is an... (read 793 more words →)
I see two major challenges (one of which leans heavily on progress in linguistics). I can see there being mathematical theory to guide candidate model decompositions (Challenge 1), but I imagine that linking up a potential model decomposition to a theory of 'semantic interpretability' (Challenge 2) is equally hard, if not harder.
Any ideas on how you plan to address Challenge 2? Maybe the most robust approach would involve active learning of the pseudocode, where a human guides the algorithm in its decomposition and labeling of each abstract computation.