
arXiv: 2304.09276
Over the last decades, deep neural networks based-models became the dominant paradigm in machine learning. Further, the use of artificial neural networks in symbolic learning has been seen as increasingly relevant recently. To study the capabilities of neural networks in the symbolic AI domain, researchers have explored the ability of deep neural networks to learn mathematical constructions, such as addition and multiplication, logic inference, such as theorem provers, and even the execution of computer programs. The latter is known to be too complex a task for neural networks. Therefore, the results were not always successful, and often required the introduction of biased elements in the learning process, in addition to restricting the scope of possible programs to be executed. In this work, we will analyze the ability of neural networks to learn how to execute programs as a whole. To do so, we propose a different approach. Instead of using an imperative programming language, with complex structures, we use the Lambda Calculus (λ-Calculus), a simple, but Turing-Complete mathematical formalism, which serves as the basis for modern functional programming languages and is at the heart of computability theory. We will introduce the use of integrated neural learning and lambda calculi formalization. Finally, we explore execution of a program in λ-Calculus is based on reductions, we will show that it is enough to learn how to perform these reductions so that we can execute any program. Keywords: Machine Learning, Lambda Calculus, Neurosymbolic AI, Neural Networks, Transformer Model, Sequence-to-Sequence Models, Computational Models
Keywords: Machine Learning, Lambda Calculus, Neurosymbolic AI, Neural Networks, Transformer Model, Sequence-to-Sequence Models, Computational Models
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Logic in Computer Science, Computer Science - Computation and Language, D.1.1, Computer Science - Artificial Intelligence, I.2.6, F.1, Machine Learning (cs.LG), Logic in Computer Science (cs.LO), Artificial Intelligence (cs.AI), I.2; I.2.6; F.1; F.1.1; D.1.1, F.1.1, Computation and Language (cs.CL)
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Logic in Computer Science, Computer Science - Computation and Language, D.1.1, Computer Science - Artificial Intelligence, I.2.6, F.1, Machine Learning (cs.LG), Logic in Computer Science (cs.LO), Artificial Intelligence (cs.AI), I.2; I.2.6; F.1; F.1.1; D.1.1, F.1.1, Computation and Language (cs.CL)
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
