Jürgen Schmidhuber
Jürgen Schmidhuber (born 17 January 1963 in Munich) is a computer scientist and artist known for his work on machine learning, Artificial Intelligence (AI), artificial neural networks, digital physics, and low-complexity art. His contributions also include generalizations of Kolmogorov complexity and the Speed Prior. From 2004 to 2009 he was professor of Cognitive Robotics at the Tech. University Munich. Since 1995 he has been co-director of the Swiss AI Lab IDSIA in Lugano, since 2009 also professor of Artificial Intelligence at the University of Lugano. Between 2009 and 2012, the recurrent neural networks and deep feedforward neural networks developed in his research group have won eight international competitions in pattern recognition and machine learning.[1] In honor of his achievements he was elected to the European Academy of Sciences and Arts in 2008.
Contributions
Recurrent neural networks
The dynamic recurrent neural networks developed in his lab are simplified mathematical models of the biological neural networks found in human brains. A particularly successful model of this type is called Long short term memory.[2] From training sequences it learns to solve numerous tasks unsolvable by previous such models. Applications range from automatic music composition to speech recognition, reinforcement learning and robotics in partially observable environments. As of 2010, his group has the best results on benchmarks in automatic handwriting recognition, obtained with deep neural networks[3] and recurrent neural networks.[4]
Artificial evolution / genetic programming
As an undergrad at TUM Schmidhuber evolved computer programs through genetic algorithms. The method was published in 1987 as one of the first papers in the emerging field that later became known as genetic programming. In the same year he published the first work on Meta-genetic programming. Since then he has co-authored numerous additional papers on artificial evolution. Applications include robot control, soccer learning, drag minimization, and time series prediction. He received several best paper awards at scientific conferences on evolutionary computation.
Neural economy
In 1989 he created the first learning algorithm for neural networks based on principles of the market economy (inspired by John Holland's bucket brigade algorithm for classifier systems): adaptive neurons compete for being active in response to certain input patterns; those that are active when there is external reward get stronger synapses, but active neurons have to pay those that activated them, by transferring parts of their synapse strengths, thus rewarding "hidden" neurons setting the stage for later success.[5]
Artificial curiosity and creativity
In 1990 he published the first in a long series of papers on artificial curiosity and creativity for an autonomous agent. The agent is equipped with an adaptive predictor trying to predict future events from the history of previous events and actions. A reward-maximizing, reinforcement learning, adaptive controller is steering the agent and gets curiosity reward for executing action sequences that improve the predictor. This discourages it from executing actions leading to boring outcomes that are either predictable or totally unpredictable.[6] Instead the controller is motivated to learn actions that help the predictor to learn new, previously unknown regularities in its environment, thus improving its model of the world, which in turn can greatly help to solve externally given tasks. This has become an important concept of developmental robotics. Schmidhuber argues that his corresponding formal theory of creativity explains essential aspects of art, science, music, and humor.[7]
Unsupervised learning / factorial codes
During the early 1990s Schmidhuber also invented a neural method for nonlinear independent component analysis (ICA) called predictability minimization. It is based on co-evolution of adaptive predictors and initially random, adaptive feature detectors processing input patterns from the environment. For each detector there is a predictor trying to predict its current value from the values of neighboring detectors, while each detector is simultaneously trying to become as unpredictable as possible.[8] It can be shown that the best the detectors can do is to create a factorial code of the environment, that is, a code that conveys all the information about the inputs such that the code components are statistically independent, which is desirable for many pattern recognition applications.
Kolmogorov complexity / computer-generated universe
In 1997 Schmidhuber published a paper based on Konrad Zuse´s assumption (1967) that the history of the universe is computable. He pointed out that the simplest explanation of the universe would be a very simple Turing machine programmed to systematically execute all possible programs computing all possible histories for all types of computable physical laws.[9][10] He also pointed out that there is an optimally efficient way of computing all computable universes based on Leonid Levin´s universal search algorithm (1973). In 2000 he expanded this work by combining Ray Solomonoff´s theory of inductive inference with the assumption that quickly computable universes are more likely than others.[11] This work on digital physics also led to limit-computable generalizations of algorithmic information or Kolmogorov complexity and the concept of Super Omegas, which are limit-computable numbers that are even more random (in a certain sense) than Gregory Chaitin´s number of wisdom Omega.[12]
Universal AI
Important research topics of his group include universal learning algorithms and universal AI[13][14] (see Gödel machine). Contributions include the first theoretically optimal decision makers living in environments obeying arbitrary unknown but computable probabilistic laws, and mathematically sound general problem solvers such as the remarkable asymptotically fastest algorithm for all well-defined problems, by his former postdoc Marcus Hutter. Based on the theoretical results obtained in the early 2000s, Schmidhuber is actively promoting the view that in the new millennium the field of general AI has matured and become a real formal science.
Low-complexity art / theory of beauty
Schmidhuber's low-complexity artworks (since 1997) can be described by very short computer programs containing very few bits of information, and reflect his formal theory of beauty[15] based on the concepts of Kolmogorov complexity and minimum description length.
Schmidhuber writes that since age 15 or so his main scientific ambition has been to build an optimal scientist, then retire. First he wants to build a scientist better than himself (he quips that his colleagues claim that should be easy) who will then do the remaining work. He claims he "cannot see any more efficient way of using and multiplying the little creativity he's got".
References
- ↑ 2012 Kurzweil AI Interview with Jürgen Schmidhuber on the eight competitions won by his Deep Learning team 2009-2012
- ↑ S. Hochreiter and J. Schmidhuber. Long Short-Term Memory. Neural Computation, 9(8):1735–1780, 1997.
- ↑ D. C. Ciresan, U. Meier, L. M. Gambardella, J. Schmidhuber. Deep Big Simple Neural Nets For Handwritten Digit Recognition. Neural Computation 22(12): 3207-3220.
- ↑ A. Graves, M. Liwicki, S. Fernandez, R. Bertolami, H. Bunke, J. Schmidhuber. A Novel Connectionist System for Improved Unconstrained Handwriting Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 5, 2009.
- ↑ J. Schmidhuber. A local learning algorithm for dynamic feedforward and recurrent networks. Connection Science, 1(4):403–412, 1989
- ↑ J. Schmidhuber. Curious model-building control systems. In Proc. International Joint Conference on Neural Networks, Singapore, volume 2, pages 1458–1463. IEEE, 1991
- ↑ J. Schmidhuber. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010). IEEE Transactions on Autonomous Mental Development, 2(3):230–247, 2010.
- ↑ J. Schmidhuber. Learning factorial codes by predictability minimization. Neural Computation, 4(6):863–879, 1992
- ↑ J. Schmidhuber. A computer scientist's view of life, the universe, and everything. Foundations of Computer Science: Potential – Theory – Cognition, Lecture Notes in Computer Science, pages 201–208, Springer, 1997
- ↑ Brian Greene, Chapter 10 of: The Hidden Reality: Parallel Universes and the Deep Laws of the Cosmos, Knopf, 2011
- ↑ J. Schmidhuber. The Speed Prior: A New Simplicity Measure Yielding Near-Optimal Computable Predictions. Proceedings of the 15th Annual Conference on Computational Learning Theory (COLT 2002), Sydney, Australia, LNAI, 216–228, Springer, 2002
- ↑ J. Schmidhuber. Hierarchies of generalized Kolmogorov complexities and nonenumerable universal measures computable in the limit. International Journal of Foundations of Computer Science 13(4):587–612, 2002
- ↑ J. Schmidhuber. Ultimate Cognition à la Gödel. Cognitive Computation 1(2):177–193, 2009
- ↑ J. Schmidhuber. Optimal Ordered Problem Solver. Machine Learning, 54, 211–254, 2004
- ↑ J. Schmidhuber. Low-Complexity Art. Leonardo, Journal of the International Society for the Arts, Sciences, and Technology, 30(2):97–103, MIT Press, 1997
External references and sources
- Google Scholar: Numerous scientific articles referencing Schmidhuber's work
- Scholarpedia article on Universal Search, discussing Schmidhuber's Speed Prior, Optimal Ordered Problem Solver, Gödel machine
- German article on Schmidhuber in CIO magazine: "Der ideale Wissenschaftler" (the ideal scientist)
- Build An Optimal Scientist, Then Retire: Interview with J. Schmidhuber in H+ magazine, 2010
- Video of Schmidhuber's talk on artificial curiosity and creativity at the Singularity Summit 2009, NYC
- TV clip: Schmidhuber on computable universes on Through the Wormhole with Morgan Freeman.
- On-going research on the iCub humanoid at the IDSIA Robotics Lab
External links
|