Schmidhuber’s Consciousness
At the 2016 AI conference in New York, Jürgen Schmidhuber introduced intriguing concepts such as Artificial Consciousness, World Model, Predictive Coding, and Data Compression. He claims that artificial systems with consciousness already exist in his lab, and consciousness arises as a byproduct of data compression during problem-solving. Let’s delve into his main ideas.
1. The Foundation of Artificial Consciousness: Data Compression
Schmidhuber explains consciousness as a byproduct of data compression. According to his theory, the ability of artificial systems to efficiently compress data during problem-solving forms the basis of consciousness. This is similar to how the human brain processes and stores complex information.
2. General-Purpose Computer and Recurrent Neural Networks
General-purpose computers, especially recurrent neural networks, are optimized for communication costs with many nodes and short wires connecting processors. These computers gather various information from the world to achieve goals or optimize rewards. Recurrent neural networks play a crucial role in this process.
3. World Model and Predictive Coding
Schmidhuber proposes a system that solves reinforcement learning and planning based on two recurrent neural networks: the controller and the world model. The controller generates actions, while the world model compresses data by making predictions about the world. All physical laws and scientific principles are compressed into sub-networks of the recurrent neural network, which are utilized by the controller.
4. Self-awareness and Artificial Curiosity
In the process of interacting with the world, artificial systems also compress their own states. This forms the basis for self-awareness. Additionally, Schmidhuber introduced artificial curiosity through Generative Adversarial Networks (GANs), presenting agents that learn via un/self-supervised learning. This is based on a minimax game where the objective function of one neural network is minimized by another.
5. Neural Networks as Programs
Schmidhuber argues that the weights of neural networks should be regarded as programs. The main goal of deep neural networks is to learn useful internal representations of observed data. The output of a neural network is differentiable concerning the program, guiding the search for better programs.
6. Predictive World Models and Self-awareness
Predictive world models that interact with the world through the controller learn to efficiently encode actions and observations through predictive coding during their lifetime. In this process, the world model forms a hierarchy of features, identifying anomalies or generating prototype encodings from the given data structure.
Conclusion
Jürgen Schmidhuber has been researching machines with artificial consciousness and emotions for 30 years. His research suggests that artificial systems can form self-awareness, interact with the world, and ultimately possess emotions and consciousness through data compression, recurrent neural networks, world models, and predictive coding. This research provides a new perspective on the future of artificial intelligence and contributes to the development of more advanced artificial systems.
Reflection Considering that the importance of predictive coding and un/self-supervised learning began to be recognized in the 2010s, it is remarkable that he identified the importance of these concepts and applied them to various cases as early as the early 1990s. Notably, Anil Seth’s 2010s theory in the field of consciousness argues that predictive coding of interoceptive information forms the basis of consciousness and self-awareness. This can relate to the compressed state of self in Schmidhuber’s world model. Additionally, the perspective of data compression aligns with Samuel Gershman’s second principle of human cognition, the principle of approximation, recently presented in his book. Data compression is not always possible, but Schmidhuber’s assertion that it forms hierarchies in recursive neural networks is deeply related.
According to his 2003 follow-up research, he introduces a formal definition of consciousness. If consciousness is defined as the ability for unlimited self-inspection and self-modification, the Gödel Machine provides a technical justification for the consciousness of artificial intelligence. By enabling limited self-inspection and self-modification, these machines offer a framework for general problem solvers capable of autonomously improving and optimizing their behavior. This achieves a form of artificial intelligence consciousness based on mathematical proofs and formal reasoning.
His information-theoretic and formal-theoretic stance and research methodology are compelling, with clarity and deep insight. However, considering the biological complexity and functions emphasized in psychology, it is necessary to discover and understand the connections with computational neuroscience and philosophy. For example, a clear explanation is needed on how primordial consciousness, as argued by affective neuroscience, can induce feelings and emotions in agents with Schmidhuber’s world model. Furthermore, in-depth research is required on the relationships between his theory and interoceptive inference, self, attention and consciousness, metacognition, and the Gödel Machine.
While Schmidhuber’s research provides clear and deep insights from an information-theoretic and formal-theoretic perspective, it is necessary to further explore its connections with the biological complexity and functions emphasized in psychology. For example, a clear explanation is needed on how primordial consciousness, as argued by affective neuroscience, can induce feelings and emotions in agents with Schmidhuber’s world model. Furthermore, in-depth research is required on the relationships between his theory and interoceptive inference, self, attention and consciousness, metacognition, and the Gödel Machine.
Summary
At the 2016 AI conference, Jürgen Schmidhuber introduced concepts of artificial consciousness, world model, predictive coding, and data compression, explaining consciousness as a byproduct of data compression during problem-solving. He proposed systems based on general-purpose computers and recurrent neural networks that gather information from the world to achieve goals and optimize rewards, using two recurrent neural networks, the controller, and the world model, for reinforcement learning and planning. These systems efficiently compress data and form self-awareness, considering the weights of neural networks as programs to guide the search for better programs. Notably, Schmidhuber recognized the importance of predictive coding and un/self-supervised learning as early as the early 1990s, connecting with Anil Seth’s interoceptive information predictive coding theory and Samuel Gershman’s data compression principle. His 2003 research introduced a formal definition of consciousness, proposing the Gödel Machine as a technical justification for AI consciousness, offering a framework for general problem solvers capable of autonomously improving their behavior. His research provides new insights into the future of AI and contributes to the development of more advanced artificial systems.
Reference
[1] Jürgen Schmidhuber’s Solution to AI Consciousness
[2] 1990: Planning & Reinforcement Learning with Recurrent World Models and Artificial Curiosity
[3] 1991: First very deep learning with unsupervised pre-training
[4] Godel Machines: Towards a Technical Justification of Consciousness
[5] X
Comments