Submitted by currentscurrents t3_10adz19 in MachineLearning
Paper: https://arxiv.org/abs/2205.01128
TL;DR It's a paper that tries to design systems that generalize. They argue there are two forms of computing: Compositional and Continuous.
Continuous computation is what neural networks are traditionally good at - creating a function that approximates a solution to a problem. Compositional computation is directly manipulating symbols, logic, ideas, etc - and unlike continuous computation, it's capable of generalizing from small datasets. But so far it's only useful inside carefully-constructed formal systems.
The authors believe research should be focused on combining the two, and implementing Compositionality fully with neural networks. They suggest some ways to do this. They also believe that the success of architectures like CNNs and Transformers comes from implementing a limited form of Compositionality.
This is a very interesting idea, but I have a little bit of skeptism:
-
This paper is heavy on theory and less so on practice. Has any followup work in this direction produced measurable results?
-
The lead author seems to have been saying things like this for a while. Sometimes older researchers have pet theories that are not broadly accepted in the field. What do other researchers think about this?
Thoughts?
navillusr t1_j43zaqk wrote
I think this is a very common belief. Symbolic systems can do many things that neural networks struggle with very sample efficiently. But they’ve failed to scale with more data as well as neural networks for most tasks, and are harder to train. If we could magically combine the reasoning ability of symbolic systems with the pattern recognition and generalization of neural networks, we would be getting very close to AGI imo. That being said idk much about recent research in symbolic reasoning so my knowledge might be outdated.