Connectionists: Scientific Integrity, the 2021 Turing Lecture, etc.

Schmidhuber Juergen juergen at idsia.ch
Tue Jan 25 12:03:24 EST 2022


PS: Terry, you also wrote: "Our precious time is better spent moving the field forward.” However, it seems like in recent years much of your own precious time has gone to promulgating a revisionist history of deep learning (and writing the corresponding "amicus curiae" letters to award committees). For a recent example, your 2020 deep learning survey in PNAS [S20] claims that your 1985 Boltzmann machine [BM] was the first NN to learn internal representations. This paper [BM] neither cited the internal representations learnt by Ivakhnenko & Lapa's deep nets in 1965 [DEEP1-2] nor those learnt by Amari’s stochastic gradient descent for MLPs in 1967-1968 [GD1-2]. Nor did your recent survey [S20] attempt to correct this as good science should strive to do. On the other hand, it seems you celebrated your co-author's birthday in a special session while you were head of NeurIPS, instead of correcting these inaccuracies and celebrating the true pioneers of deep learning, such as Ivakhnenko and Amari. Even your recent interview https://blog.paperspace.com/terry-sejnowski-boltzmann-machines/ claims: "Our goal was to try to take a network with multiple layers - an input layer, an output layer and layers in between – and make it learn. It was generally thought, because of early work that was done in AI in the 60s, that no one would ever find such a learning algorithm because it was just too mathematically difficult.” You wrote this although you knew exactly that such learning algorithms were first created in the 1960s, and that they worked. You are a well-known scientist, head of NeurIPS, and chief editor of a major journal. You must correct this. We must all be better than this as scientists. We owe it to both the past, present, and future scientists as well as those we ultimately serve.

The last paragraph of my report https://people.idsia.ch/~juergen/scientific-integrity-turing-award-deep-learning.html quotes Elvis Presley: "Truth is like the sun. You can shut it out for a time, but it ain't goin' away.” I wonder how the future will reflect on the choices we make now.      

Jürgen


> On 3 Jan 2022, at 11:38, Schmidhuber Juergen <juergen at idsia.ch> wrote:
> 
> Terry, please don't throw smoke candles like that!  
> 
> This is not about basic math such as Calculus (actually first published by Leibniz; later Newton was also credited for his unpublished work; Archimedes already had special cases thereof over 2000 years ago; the Indian Kerala school made essential contributions around 1400). In fact, my report addresses such smoke candles in Sec. XII: "Some claim that 'backpropagation' is just the chain rule of Leibniz (1676) & L'Hopital (1696).' No, it is the efficient way of applying the chain rule to big networks with differentiable nodes (there are also many inefficient ways of doing this). It was not published until 1970 [BP1]."
> 
> You write: "All these threads will be sorted out by historians one hundred years from now." To answer that, let me just cut and paste the last sentence of my conclusions: "However, today's scientists won't have to wait for AI historians to establish proper credit assignment. It is easy enough to do the right thing right now."
> 
> You write: "let us be good role models and mentors" to the new generation. Then please do what's right! Your recent survey [S20] does not help. It's mentioned in my report as follows: "ACM seems to be influenced by a misleading 'history of deep learning' propagated by LBH & co-authors, e.g., Sejnowski [S20] (see Sec. XIII). It goes more or less like this: 'In 1969, Minsky & Papert [M69] showed that shallow NNs without hidden layers are very limited and the field was abandoned until a new generation of neural network researchers took a fresh look at the problem in the 1980s [S20].' However, as mentioned above, the 1969 book [M69] addressed a 'problem' of Gauss & Legendre's shallow learning (~1800)[DL1-2] that had already been solved 4 years prior by Ivakhnenko & Lapa's popular deep learning method [DEEP1-2][DL2] (and then also by Amari's SGD for MLPs [GD1-2]). Minsky was apparently unaware of this and failed to correct it later [HIN](Sec. I).... deep learning research was alive and kicking also in the 1970s, especially outside of the Anglosphere."
> 
> Just follow ACM's Code of Ethics and Professional Conduct [ACM18] which states: "Computing professionals should therefore credit the creators of ideas, inventions, work, and artifacts, and respect copyrights, patents, trade secrets, license agreements, and other methods of protecting authors' works." No need to wait for 100 years. 
> 
> Jürgen
> 
> 
> 
> 
> 
>> On 2 Jan 2022, at 23:29, Terry Sejnowski <terry at snl.salk.edu> wrote:
>> 
>> We would be remiss not to acknowledge that backprop would not be possible without the calculus,
>> so Isaac newton should also have been given credit, at least as much credit as Gauss.
>> 
>> All these threads will be sorted out by historians one hundred years from now.
>> Our precious time is better spent moving the field forward.  There is much more to discover.
>> 
>> A new generation with better computational and mathematical tools than we had back
>> in the last century have joined us, so let us be good role models and mentors to them.
>> 
>> Terry
>> 
>> -----
>> 
>> On 1/2/2022 5:43 AM, Schmidhuber Juergen wrote:
>>> Asim wrote: "In fairness to Jeffrey Hinton, he did acknowledge the work of Amari in a debate about connectionism at the ICNN’97 .... He literally said 'Amari invented back propagation'..." when he sat next to Amari and  Werbos. Later, however, he failed to cite Amari’s stochastic gradient descent (SGD) for multilayer NNs (1967-68) [GD1-2a] in his 2015 survey [DL3], his 2021 ACM lecture [DL3a], and other surveys.  Furthermore, SGD [STO51-52] (Robbins, Monro, Kiefer, Wolfowitz, 1951-52) is not even backprop. Backprop is just a particularly efficient way of computing gradients in differentiable networks, known as the reverse mode of automatic differentiation, due to Linnainmaa (1970) [BP1] (see also Kelley's precursor of 1960 [BPa]). Hinton did not cite these papers either, and in 2019 embarrassingly did not hesitate to accept an award for having "created ... the backpropagation algorithm” [HIN]. All references and more on this can be found in the report, especially in Sec. XII.
>>> 
>>> The deontology of science requires: If one "re-invents" something that was already known, and only becomes aware of it later, one must at least clarify it later [DLC], and correctly give credit in all follow-up papers and presentations. Also, ACM's Code of Ethics and Professional Conduct [ACM18] states: "Computing professionals should therefore credit the creators of ideas, inventions, work, and artifacts, and respect copyrights, patents, trade secrets, license agreements, and other methods of protecting authors' works." LBH didn't.
>>> 
>>> Steve still doesn't believe that linear regression of 200 years ago is equivalent to linear NNs. In a mature field such as math we would not have such a discussion. The math is clear. And even today, many students are taught NNs like this: let's start with a linear single-layer NN (activation = sum of weighted inputs). Now minimize mean squared error on the training set. That's good old linear regression (method of least squares). Now let's introduce multiple layers and nonlinear but differentiable activation functions, and derive backprop for deeper nets in 1960-70 style (still used today, half a century later).
>>> 
>>> Sure, an important new variation of the 1950s (emphasized by Steve) was to transform linear NNs into binary classifiers with threshold functions. Nevertheless, the first adaptive NNs (still widely used today) are 1.5 centuries older except for the name.
>>> 
>>> Happy New Year!
>>> 
>>> Jürgen
>>> 
>>> 
>>>> On 2 Jan 2022, at 03:43, Asim Roy <ASIM.ROY at asu.edu> wrote:
>>>> 
>>>> And, by the way, Paul Werbos was also there at the same debate. And so was Teuvo Kohonen.
>>>> 
>>>> Asim
>>>> 
>>>> -----Original Message-----
>>>> From: Asim Roy
>>>> Sent: Saturday, January 1, 2022 3:19 PM
>>>> To: Schmidhuber Juergen <juergen at idsia.ch>; connectionists at cs.cmu.edu
>>>> Subject: RE: Connectionists: Scientific Integrity, the 2021 Turing Lecture, etc.
>>>> 
>>>> In fairness to Jeffrey Hinton, he did acknowledge the work of Amari in a debate about connectionism at the ICNN’97 (International Conference on Neural Networks) in Houston. He literally said "Amari invented back propagation" and Amari was sitting next to him. I still have a recording of that debate.
>>>> 
>>>> Asim Roy
>>>> Professor, Information Systems
>>>> Arizona State University
>>>> https://isearch.asu.edu/profile/9973
>>>> https://lifeboat.com/ex/bios.asim.roy
>>> 
>>> On 2 Jan 2022, at 02:31, Stephen José Hanson <jose at rubic.rutgers.edu> wrote:
>>> 
>>> Juergen:  Happy New Year!
>>> 
>>> "are not quite the same"..
>>> 
>>> I understand that its expedient sometimes to use linear regression to approximate the Perceptron.(i've had other connectionist friends tell me the same thing) which has its own incremental update rule..that is doing <0,1> classification.    So I guess if you don't like the analogy to logistic regression.. maybe Fisher's LDA?  This whole thing still doesn't scan for me.
>>> 
>>> So, again the point here is context.   Do you really believe that Frank Rosenblatt didn't reference Gauss/Legendre/Laplace  because it slipped his mind??   He certainly understood modern statistics (of the 1940s and 1950s)
>>> 
>>> Certainly you'd agree that FR could have referenced linear regression as a precursor, or "pretty similar" to what he was working on, it seems disingenuous to imply he was plagiarizing Gauss et al.--right?  Why would he?
>>> 
>>> Finally then, in any historical reconstruction, I can think of,  it just doesn't make sense.    Sorry.
>>> 
>>> Steve
>>> 
>>> 
>>>> -----Original Message-----
>>>> From: Connectionists <connectionists-bounces at mailman.srv.cs.cmu.edu> On Behalf Of Schmidhuber Juergen
>>>> Sent: Friday, December 31, 2021 11:00 AM
>>>> To: connectionists at cs.cmu.edu
>>>> Subject: Re: Connectionists: Scientific Integrity, the 2021 Turing Lecture, etc.
>>>> 
>>>> Sure, Steve, perceptron/Adaline/other similar methods of the 1950s/60s are not quite the same, but the obvious origin and ancestor of all those single-layer  “shallow learning” architectures/methods is indeed linear regression; today’s simplest NNs minimizing mean squared error are exactly what they had 2 centuries ago. And the first working deep learning methods of the 1960s did NOT really require “modern” backprop (published in 1970 by Linnainmaa [BP1-5]). For example, Ivakhnenko & Lapa (1965) [DEEP1-2] incrementally trained and pruned their deep networks layer by layer to learn internal representations, using regression and a separate validation set. Amari (1967-68)[GD1] used stochastic gradient descent [STO51-52] to learn internal representations WITHOUT “modern" backprop in his multilayer perceptrons. Jürgen
>>>> 
>>>> 
>>>>> On 31 Dec 2021, at 18:24, Stephen José Hanson <jose at rubic.rutgers.edu> wrote:
>>>>> 
>>>>> Well the perceptron is closer to logistic regression... but the heaviside function  of course is <0,1>   so technically not related to linear regression which is using covariance to estimate betas...
>>>>> 
>>>>> does that matter?  Yes, if you want to be hyper correct--as this appears to be-- Berkson (1944) coined the logit.. as log odds.. for probabilistic classification.. this was formally developed by Cox in the early 60s, so unlikely even in this case to be a precursor to perceptron.
>>>>> 
>>>>> My point was that DL requires both Learning algorithm (BP) and an
>>>>> architecture.. which seems to me much more responsible for the the success of Dl.
>>>>> 
>>>>> S
>>>>> 
>>>>> 
>>>>> 
>>>>> On 12/31/21 4:03 AM, Schmidhuber Juergen wrote:
>>>>>> Steve, this is not about machine learning in general, just about deep
>>>>>> learning vs shallow learning. However, I added the Pandemonium -
>>>>>> thanks for that! You ask: how is a linear regressor of 1800
>>>>>> (Gauss/Legendre) related to a linear neural network? It's formally
>>>>>> equivalent, of course! (The only difference is that the weights are
>>>>>> often called beta_i rather than w_i.) Shallow learning: one adaptive
>>>>>> layer. Deep learning: many adaptive layers. Cheers, Jürgen
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> On 31 Dec 2021, at 00:28, Stephen José Hanson
>>>>>>> <jose at rubic.rutgers.edu>
>>>>>>> wrote:
>>>>>>> 
>>>>>>> Despite the comprehensive feel of this it still appears to me to be  too focused on Back-propagation per se.. (except for that pesky Gauss/Legendre ref--which still baffles me at least how this is related to a "neural network"), and at the same time it appears to be missing other more general epoch-conceptually relevant cases, say:
>>>>>>> 
>>>>>>> Oliver Selfridge  and his Pandemonium model.. which was a hierarchical feature analysis system.. which certainly was in the air during the Neural network learning heyday...in fact, Minsky cites Selfridge as one of his mentors.
>>>>>>> 
>>>>>>> Arthur Samuels:  Checker playing system.. which learned a evaluation function from a hierarchical search.
>>>>>>> 
>>>>>>> Rosenblatt's advisor was Egon Brunswick.. who was a gestalt perceptual psychologist who introduced the concept that the world was stochastic and the the organism had to adapt to this variance somehow.. he called it "probabilistic functionalism"  which brought attention to learning, perception and decision theory, certainly all piece parts of what we call neural networks.
>>>>>>> 
>>>>>>> There are many other such examples that influenced or provided context for the yeasty mix that was 1940s and 1950s where Neural Networks  first appeared partly due to PItts and McCulloch which entangled the human brain with computation and early computers themselves.
>>>>>>> 
>>>>>>> I just don't see this as didactic, in the sense of a conceptual view of the  multidimensional history of the         field, as opposed to  a 1-dimensional exegesis of mathematical threads through various statistical algorithms.
>>>>>>> 
>>>>>>> Steve
>>>>>>> 
>>>>>>> On 12/30/21 1:03 PM, Schmidhuber Juergen wrote:
>>>>>>> 
>>>>>>>> Dear connectionists,
>>>>>>>> 
>>>>>>>> in the wake of massive open online peer review, public comments on the connectionists mailing list [CONN21] and many additional private comments (some by well-known deep learning pioneers) helped to update and improve upon version 1 of the report. The essential statements of the text remain unchanged as their accuracy remains unchallenged. I'd like to thank everyone from the bottom of my heart for their feedback up until this point and hope everyone will be satisfied with the changes. Here is the revised version 2 with over 300 references:
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> https://urldefense.com/v3/__https://people.idsia.ch/*juergen/scient
>>>>>>>> ific-integrity-turing-award-deep-learning.html__;fg!!IKRxdwAv5BmarQ
>>>>>>>> !NsJ4lf4yO2BDIBzlUVfGKvTtf_QXY8dpZaHzCSzHCvEhXGJUTyRTzZybDQg-DZY$
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> In particular, Sec. II has become a brief history of deep learning up to the 1970s:
>>>>>>>> 
>>>>>>>> Some of the most powerful NN architectures (i.e., recurrent NNs) were discussed in 1943 by McCulloch and Pitts [MC43] and formally analyzed in 1956 by Kleene [K56] - the closely related prior work in physics by Lenz, Ising, Kramers, and Wannier dates back to the 1920s [L20][I25][K41][W45]. In 1948, Turing wrote up ideas related to artificial evolution [TUR1] and learning NNs. He failed to formally publish his ideas though, which explains the obscurity of his thoughts here. Minsky's simple neural SNARC computer dates back to 1951. Rosenblatt's perceptron with a single adaptive layer learned in 1958 [R58] (Joseph [R61] mentions an earlier perceptron-like device by Farley & Clark); Widrow & Hoff's similar Adaline learned in 1962 [WID62]. Such single-layer "shallow learning" actually started around 1800 when Gauss & Legendre introduced linear regression and the method of least squares [DL1-2] - a famous early example of pattern recognition and generalization from training data through a parameterized predictor is Gauss' rediscovery of the asteroid Ceres based on previous astronomical observations. Deeper multilayer perceptrons (MLPs) were discussed by Steinbuch [ST61-95] (1961), Joseph [R61] (1961), and Rosenblatt [R62] (1962), who wrote about "back-propagating errors" in an MLP with a hidden layer [R62], but did not yet have a general deep learning algorithm for deep MLPs  (what's now called backpropagation is quite different and was first published by Linnainmaa in 1970 [BP1-BP5][BPA-C]). Successful learning in deep architectures started in 1965 when Ivakhnenko & Lapa published the first general, working learning algorithms for deep MLPs with arbitrarily many hidden layers (already containing the now popular multiplicative gates) [DEEP1-2][DL1-2]. A paper of 1971 [DEEP2] already described a deep learning net with 8 layers, trained by their highly cited method which was still popular in the new millennium [DL2], especially in Eastern Europe, where much of Machine Learning was born [MIR](Sec. 1)[R8]. LBH fai
led to cite this, just like they failed to cite Amari [GD1], who in 1967 proposed stochastic gradient descent [STO51-52] (SGD) for MLPs and whose implementation [GD2,GD2a] (with Saito) learned internal representations at a time when compute was billions of times more expensive than today (see also Tsypkin's work [GDa-b]). (In 1972, Amari also published what was later sometimes called the Hopfield network or Amari-Hopfield Network [AMH1-3].) Fukushima's now widely used deep convolutional NN architecture was first introduced in the 1970s [CNN1].
>>>>>> 
>>>>>>>> Jürgen
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> ******************************
>>>>>>>> 
>>>>>>>> On 27 Oct 2021, at 10:52, Schmidhuber Juergen
>>>>>>>> 
>>>>>>>> <juergen at idsia.ch>
>>>>>>>> 
>>>>>>>> wrote:
>>>>>>>> 
>>>>>>>> Hi, fellow artificial neural network enthusiasts!
>>>>>>>> 
>>>>>>>> The connectionists mailing list is perhaps the oldest mailing list on ANNs, and many neural net pioneers are still subscribed to it. I am hoping that some of them - as well as their contemporaries - might be able to provide additional valuable insights into the history of the field.
>>>>>>>> 
>>>>>>>> Following the great success of massive open online peer review
>>>>>>>> (MOOR) for my 2015 survey of deep learning (now the most cited
>>>>>>>> article ever published in the journal Neural Networks), I've
>>>>>>>> decided to put forward another piece for MOOR. I want to thank the
>>>>>>>> many experts who have already provided me with comments on it.
>>>>>>>> Please send additional relevant references and suggestions for
>>>>>>>> improvements for the following draft directly to me at
>>>>>>>> 
>>>>>>>> juergen at idsia.ch
>>>>>>>> 
>>>>>>>> :
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> https://urldefense.com/v3/__https://people.idsia.ch/*juergen/scient
>>>>>>>> ific-integrity-turing-award-deep-learning.html__;fg!!IKRxdwAv5BmarQ
>>>>>>>> !NsJ4lf4yO2BDIBzlUVfGKvTtf_QXY8dpZaHzCSzHCvEhXGJUTyRTzZybDQg-DZY$
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> The above is a point-for-point critique of factual errors in ACM's justification of the ACM A. M. Turing Award for deep learning and a critique of the Turing Lecture published by ACM in July 2021. This work can also be seen as a short history of deep learning, at least as far as ACM's errors and the Turing Lecture are concerned.
>>>>>>>> 
>>>>>>>> I know that some view this as a controversial topic. However, it is the very nature of science to resolve controversies through facts. Credit assignment is as core to scientific history as it is to machine learning. My aim is to ensure that the true history of our field is preserved for posterity.
>>>>>>>> 
>>>>>>>> Thank you all in advance for your help!
>>>>>>>> 
>>>>>>>> Jürgen Schmidhuber
>>>>>>>> 
>>>>>>>> 




More information about the Connectionists mailing list