Submitted by currentscurrents t3_10adz19 in MachineLearning
Paper: https://arxiv.org/abs/2205.01128
TL;DR It's a paper that tries to design systems that generalize. They argue there are two forms of computing: Compositional and Continuous.
Continuous computation is what neural networks are traditionally good at - creating a function that approximates a solution to a problem. Compositional computation is directly manipulating symbols, logic, ideas, etc - and unlike continuous computation, it's capable of generalizing from small datasets. But so far it's only useful inside carefully-constructed formal systems.
The authors believe research should be focused on combining the two, and implementing Compositionality fully with neural networks. They suggest some ways to do this. They also believe that the success of architectures like CNNs and Transformers comes from implementing a limited form of Compositionality.
This is a very interesting idea, but I have a little bit of skeptism:
-
This paper is heavy on theory and less so on practice. Has any followup work in this direction produced measurable results?
-
The lead author seems to have been saying things like this for a while. Sometimes older researchers have pet theories that are not broadly accepted in the field. What do other researchers think about this?
Thoughts?
Diffeologician t1_j43qshg wrote
Isn’t that the whole point of differentiable programming?