Small History of Deep Learning:
- 1943 → McCulloch and Pitts, networks of binary neurons can do logic; if we have binary elements, can do logical computations and inference - for reasoning
- 1947 → Donald Hebb, Hebbian Synaptic Plasticity → by changing the strengths of the neurons in the brain , they can change the strength of the networks
- 1948 → Norbert Wiener, Cybernetics → Systems Theory now, Optimal Filter, Feedback,Autopoises, Auto Organization; Very complex phenomenon emerging from simple systems. Connecting lots of lots of simple things and rules → they can self organize into rules (seen in biology)
- 1957 → Frank Rosenblatt ; Perceptron → simple idea → supervised learning ; having a system that can classify → patterns by adjusting weigths → synapses on a neuron.
- 1961 → Bernie Widrow → Adaline → very similar to perceptron → neuro classifier
- 1962 → Hubel and Wiesel → Visual Cortex Architecture → architecture of the visual cortex → lead to a Nobel Prize.
- 1969 → Minsky and Papert → Limits of the Perceptron. → lead to Winter of AI → killed the whole field in the late 70s.
- 1970s → statistical pattern recognition → Duda and Hart / adpative filters
- 1979 → Kunihiko Fukushima , Neocognitron → neural nets and cognitron
- 1982 → HopField Networks → physicists ;
- 1982 → Hinton and Sejnowski → Boltzmann Machines → Physicsist
- 1985/86 → Practical Backpropogation for neural net training
- 1989 → Convolutional Networks → started the NIPS conference; and brought back the attention to neural networks.
- 1991 → Bottou and Gallinari → module based automatic differnetiation
- 1995 → Hochreiter and Schmidhuber → LSTM recurrent net
- 1996 → Structured prediction with neural nets → graph transformer nets
- 2003 → Yoshua Bengio → Neural Language Model
- 2006 → Layer-wise unsupervised pre-training of deep networks.
- 2010 → Collobert and Weston → self-supervised neural nets in NLP
- 2012 → AlexNet /convent on GPU/ Object classification
- 2015 → I. Sutskever → Neural machine translation with multilayer LSTM
- 2015 → Weston, Chopra and Bordes → Memory Networks
- 2016 → Bahdanau,Cho,Bengio → GRU, Attention Mechanism
- 2016 → Kaiming He → ResNet
Supervised Learning goes back to the Perceptron and Adaline:
-
Main thing was the McCullough-Pitts Binary Neuron where the perform weighted avergeas and compared to a threshold and turns off if less than a threshold.
-
-
Perceptron→ weights are motorized potentiometers.
-
Adaline → Weights are electrochemical “mimstors”
Hot Topics in AI:(2022 version)
- Self Supervised Learning - learning without looking into the data before → pretrain a neural net to understand the data without fine tuning for processing. Then fine tune of what thing we would want to optimize. Improved a lot of computer vision.
- Machine Based Reasoning - learning like humans → Animals and Humans can do perceptions, navigating etc that dont require reasoning → for long chains of reasoning and planning → how to modify our architectures for reasons. Machine networks → reason on it and then change the basis of the underlying networks.
- How to train systems to train on action plans and stuff ?
Limitations and Quantum Computing and Non binary states
- one big challenge, can we come up with paradigms that will allow the machines to come up with reasoning on its paradigms by observing and common sense. No machine has common sense yet ! AI systems have very narrow intelligence.
- Quantum Computing and AI is still much up in the air , and in a academic research subject it is interesting to observe.