![]() | This is an archive of past discussions about Neural network (machine learning). Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Today's deep neural networks are based on early work in statistics over 200 years ago. The simplest kind of feedforward neural network (FNN) is a linear network, which consists of a single layer of output nodes with linear activation functions; the inputs are fed directly to the outputs via a series of weights. The sum of the products of the weights and the inputs is calculated at each node. The mean squared errors between these calculated outputs and the given target values are minimized by creating an adjustment to the weights. This technique has been known for over two centuries as the method of least squares or linear regression. It was used as a means of finding a good rough linear fit to a set of points by Legendre (1805) and Gauss (1795) for the prediction of planetary movement.[1][2][3][4][5]
In 1958, psychologist Frank Rosenblatt described the perceptron, one of the first implemented artificial neural networks,[6][7][8][9] funded by the United States Office of Naval Research.[10] R. D. Joseph (1960)[11] mentions an even earlier perceptron-like device by Farley and Clark[4]: "Farley and Clark of MIT Lincoln Laboratory actually preceded Rosenblatt in the development of a perceptron-like device." However, "they dropped the subject." Farley and Clark[12] (1954) also used computational machines to simulate a Hebbian network. Other neural network computational machines were created by Rochester, Holland, Habit and Duda (1956).[13] The perceptron raised public excitement for research in Artificial Neural Networks, causing the US government to drastically increase funding. This contributed to "the Golden Age of AI" fueled by the optimistic claims made by computer scientists regarding the ability of perceptrons to emulate human intelligence.[14] The first perceptrons did not have adaptive hidden units. However, Joseph (1960)[11] also discussed multilayer perceptrons with an adaptive hidden layer. Rosenblatt (1962)[15]: section 16 cited and adopted these ideas, also crediting work by H. D. Block and B. W. Knight. Unfortunately, these early efforts did not lead to a working learning algorithm for hidden units, i.e., deep learning.
References
DLhistory
was invoked but never defined (see the help page).