All of Guillaume Corlouer's Comments + Replies

Another perspective would be too look at the activations of an autoregressive deep learning model, e.g. a transformer, during inference as a stochastic process: the collection of activation  at some layer as random variables indexed by time t, where t is token position.

One could for example look at mutual information between the history  and the future of the activations , or look at (conditional) mutual information between the past and future of subprocesses of  (note: transfer entropy can be a... (read more)

An interesting analogy, closer to ML, would be to look at neuroscience. It's an older field than ML, and it seems that the physics perspective has been fairly productive, even though not successful at providing a grand unified theory of cognition yet. Some examples: 

  • Using methods from electric circuits to explain neurons (Hodgkin-Huxley model, cable theory)
  • Dynamical systems to explain phenomena like synchronization in neuronal oscillations (ex: Kuramoto model)
  • Ising models to model some collective behaviour of neurons
  • Information theory is commonly used
... (read more)
2Lauren Greenspan
Thanks for the recommendation! The pathways of scientific progress here seem very interesting (for example: physics -> neuro -> AI -> ... v. physics -> AI -> neuro -> ...), particularly if we think about feeding back between experimental and theoretical support to build up a general understanding. Physics is really good at fitting theories together in a mosaic -- at a large scale you have a nice picture of the universe, and the tiles (theories) fit together but aren't part of the same continuous picture, allowing for some separation between different regimes of validity. It's not a perfect analogy, but it says something about physics' ability to split the difference between reductionism and emergence. It would be nice to have a similar picture in neuroscience (and AI), though this might be more difficult.
1MichaelCorey
The debate on AI’s physical nature is fascinating. While AI itself is a set of algorithms, it relies on physical infrastructure—hardware, servers, and energy. Some argue that AI "exists" within digital frameworks but lacks tangible form. The philosophical implications are worth exploring. If you're interested in writing about AI or tech advancements, GradesFixer https://gradesfixer.com/blog/best-literary-topics-for-writing-essays-that-will-be-popular-in-2025/ has a great collection of literary topics that will be trending in 2025. Their resource on best literary topics can help refine your arguments and structure your thoughts effectively.

Interesting! Perhaps one way to not be fooled by such situations could be to use a non-parametric statistical test. For example, we could apply permutation testing: by resampling the data to break its correlation structure and performing PCA on each permuted dataset, we can form a null distribution of eigenvalues. Then, by comparing the eigenvalues from the original data to this null distribution, we could assess whether the observed structure is unlikely under randomness. Specifically, we’d look at how extreme each original eigenvalue is relative to those... (read more)

Right, I got confused because I thought your problem was about trying to define a measure of optimisation power - for ex analogous to the Yudkowsky measure - that was also referring to a utility function, while being invariant from scaling and translation but this is different from asking

"what fraction of the default expected utility comes from outcomes at least as good as this one?’"

What about optimisation power of  as a measure of outcome that have utility greater than the utility of 

Let  be the set of outcome with utility greater than  according to utility function :

The set    is invariant under translation and non-zero rescaling of the utility function  and we define the optimisation power of the outcome ' according to utility function  as:

This does not suffer from compar... (read more)

2Alexander Gietelink Oldenziel
Yeah this is the expectation of the Yudkowsky measure I think?

Nice recommendations! In addition to brain enthusiasts being useful for empirical work, there also are theoretical tools from system neuroscience that could be useful for AI safety. One area in particular would be for interpretability: if we want to model a network at various levels of "emergence", recent development in information decomposition and multivariate information theory to move beyond pairwise interaction in a neural network might be very useful. Also see recent publications to model synergestic information  and dynamical independance to pe... (read more)

2Jan
Great points, thanks for the comment! :) I agree that there are potentially some very low-hanging fruits. I could even imagine that some of these methods work better in artificial networks than in biological networks (less noise, more controlled environment). But I believe one of the major bottlenecks might be that the weights and activations of an artificial neural network are just so difficult to access? Putting the weights and activations of a large model like GPT-3 under the microscope requires impressive hardware (running forward passes, storing the activations, transforming everything into a useful form, ...) and then there are so many parameters to look at.  Giving researchers structured access to the model via a research API could solve a lot of those difficulties and appears like something that totally should exist (although there is of course the danger of accelerating progress on the capabilities side also).

Thanks for all the useful links! I'm also always happy to receive more feedback.

I agree that the sense in which I use metaethics in this post is different from what academic philosophers usually call metaethics. I have the impression that metaethics, in academic sense, and metaphilosophy are somehow related. Studying what morality itself is, how to select ethical theories and what is the process behind ethical reasoning seems not independent. For example if moral nihilism is more plausible then it seems to be less likely that there is some meaningful... (read more)

Sure, I'm happy to read/discuss your ideas about this topic.

I am not sure about what computer aided analysis mean but one possibility could be to have formal ethical theories and prove some theorem inside their formal framework. But this raises questions about the sort of formal framework that one could use to 'prove theorems' under ethics in a meaningful way.

1Teerth Aloke
Till this point, I have heard the idea of an ethics axiomatic system several times. But, no suggestion of what such axioms could be. Computer aided analysis in the sense of an automated theorem checker to search for contradictions in the system.