![]() | This is an archive of past discussions about Neural network (machine learning). Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
This article fails. Wikipedia is supposed to be an encyclopedia, not a graduate math text, comprehensible only by math geeks. More plain text for normal people is sorely needed. I could not make head nor tails of this article, and I hold two degrees in computer science. —Preceding unsigned comment added by 24.18.203.194 (talk) 11:27, 8 May 2010 (UTC)
It's worse than abject failure. It's a complete mess. I would recommend anybody who comes here to run away as quickly as possible. There are several excellent books about NNs that range from popsci to university level text books. There is plenty of free online material. There are also plenty of excellent free lectures available from MIT and others. Some excellent youtube videos. Wikipedia should be ashamed of this disaster. The article should be taken down until its readable. — Preceding unsigned comment added by 2601:646:8580:19:5D95:B046:D9AD:5FA3 (talk) 22:46, 22 October 2017 (UTC)
Consensus is to not merge. NN, BNN and ANN are three separate entities. Consensus is to keep three separate articles and slim each down to a more specific version by removing NN from ANN and ANN from BNN etc.
95% of the article Neural network is about Artificial neural networks. Keeping these two articles is unnecessary forking and only supports continued divergence/overlap of the texts. The article must be merged and then restructured by according to reasonable subtopics. The page Neural network must be a disambig page for ANN, BNN and Neural Networks, the Official Journal of the International Neural Network Society, European Neural Network Society & Japanese Neural Network Society. Twri (talk) 23:32, 24 November 2008 (UTC)
OpposeChaosdruid (talk) 03:26, 12 July 2010 (UTC)
I currently am at the RoboCup 2009 Competition in graz, where I found the site http://www.dkriesel.com/en/science/robocup_2009_graz because different to the robocup site ;) it presents recent news and pictures about robocup.
What I found there might be something for this wikipage: http://www.dkriesel.com/en/science/neural_networks a neural networks PDF manuscript is presented that seems to be extended often, is free, contains whole lots of illustrations and (this is special) is available in English and German Language. I also noticed that its german version is linked in the german wikipedia. I want to start a discussion if it should be added as weblink in this article. If there will be no protest, I would try and add it in the next few days. 91.143.103.39 (talk) 07:42, 4 July 2009 (UTC)
Does anyone else feel like this page is incomprehensible? Paskari (talk) 16:38, 13 January 2009 (UTC)
Yeah, reading the article one doesn't know what all of this stuff have to do with neurons (I mean, the article apparently only talks about functions). —Preceding unsigned comment added by 80.31.182.27 (talk) 11:32, 4 March 2009 (UTC)
I prefer leaving "Neural Network" as it is because the contents on the heading "Neural Network" gives the basic understanding of the Biological Neural Network and differs, in a great way, from Artifical Neural Network and its understanding.
The main discussion in neural network is about artificial Neural Network.So they should be merged with a discussion of Natural Neural network in introduction.Bpavel88 (talk) 19:03, 1 May 2009 (UTC)
I would agree that substantial differences lie between the two types, and that there is specific terminology used for the artifical types that would not be appropriate for the non-artificial page 220.253.48.185 (talk) 03:49, 7 June 2010 (UTC)
There are a lot of common concepts in artificial and biological neural networks, but they are also quite different. We don't really know how biological neural networks operate, and at the same time artificial neural networks use methods that we don't know is apropriate for biological neural networks. I would say keep them apart for now and probably for foreseeable future. Jeblad (talk) 13:56, 31 March 2019 (UTC)
I think this page should have 2-3 paragraphs tops for all the types of neural networks. than we can split up the types into a new page, making it more readable. Oldag07 (talk) 17:21, 20 August 2009 (UTC)
I came to this page to find out about the computational power of neural networks. There was a claim that a particular neural network (not described) has universal turing power, but the link and DOI in the citation both seem to point to a non-existent paper.120.16.58.115 (talk) 04:17, 15 October 2009 (UTC)
The remarks by Dewdney are really from a sour physicist missing the point. For difficult problems you first want to see the existence proof that a universal function approximator can do (part of) the job. Once that is the case you go hunt for the concise or 'real' solution. The Dewdy comment is very surprising, because that was about six years after the invention of the convolutional neural MLP by Yann LeCun, still unbeaten in handwritten character recognition after twenty years (better than 99.3 percent on the NIST benchmark). If the citation to Dewdney remains in there, the balance requires that (more) success stories are presented more clearly in this article. — Preceding unsigned comment added by 129.125.178.72 (talk) 15:48, 3 October 2011 (UTC)
Dewdney's criticism is indeed outdated. One should add something about the spectacular recent successes since 2009: Between 2009 and 2012, the recurrent neural networks and deep feedforward neural networks developed in the research group of Jürgen Schmidhuber at the Swiss AI Lab IDSIA have won eight international competitions in pattern recognition and machine learning[1]. For example, the bi-directional and multi-dimensional Long short term memory (LSTM)[2][3] by Alex Graves et al. won three competitions in connected handwriting recognition at the 2009 International Conference on Document Analysis and Recognition (ICDAR), without any prior knowledge about the three different languages to be learned. Recent deep learning methods for feedforward networks alternate convolutional layers[4] and max-pooling layers[5], topped by several pure classification layers. Fast GPU-based implementations of this approach by Dan Ciresan and colleagues at IDSIA have won several pattern recognition contests, including the IJCNN 2011 Traffic Sign Recognition Competition[6], the ISBI 2012 Segmentation of Neuronal Structures in Electron Microscopy Stacks challenge[7], and others. Their neural networks also were the first artificial pattern recognizers to achieve human-competitive or even superhuman performance[8] on important benchmarks such as traffic sign recognition (IJCNN 2012), or the famous MNIST handwritten digits problem of Yann LeCun at NYU. Deep, highly nonlinear neural architectures similar to the 1980 Neocognitron by Kunihiko Fukushima[9] and the "standard architecture of vision"[10]
can also be pre-trained by unsupervised methods[11][12] of Geoff Hinton's lab at Toronto University.
Deeper Learning (talk) 22:23, 13 December 2012 (UTC)
References
{{cite journal}}
: Check |authorlink1=
value (help)
The entire article is absolutely terrible; there are so many good facts, but the organization is atrocious. A team needs to come in, clean up the article, word it well, and it's quite a shame because of how developed it's become. If anyone wants this article to at least reach a B-class rating on the quality scale (which is extremely important due to the article's importance in Wikipedia), we really need to clean it up. It's incomprehensible, and as someone pointed out above, it just talks about the functions of an artificial neural network, rather than how it's modelled upon biological neural networks, which is the principle purpose of this article, to explain how the two are related, and the history/applications of the system. Even worse, there are NO citations in the first few sections, and they are quite scarce. There is an excessive amount of subsections, which themselves are mere paragraphs.
Final verdict: This article needs to be re-written!
Thanks, Rifasj123 (talk) 22:47, 19 June 2012 (UTC)
“ | The original inspiration for the term Artificial Neural Network came from examination of central nervous systems and their neurons, axons, dendrites, and synapses, which constitute the processing elements of biological neural networks investigated by neuroscience. | ” |
This almost seems as if it has deliberately been written to confuse the reader. Again, going down to Models,
“ | Neural network models in artificial intelligence are usually referred to as artificial neural networks (ANNs); these are essentially simple mathematical models defining a function or a distribution over or both and , but sometimes models are also intimately associated with a particular learning algorithm or learning rule. | ” |
This just doesn't make any sense! Nobody can get anywhere reading this article, it's just babbling and jargon glued together with mumbo-jumbo. JoshuSasori (talk) 03:50, 14 September 2012 (UTC)
"Deep learning" is little more than a fad term for the current generation of neural nets, and this page describes neural net technology almost exclusively. The page neural network could do with an update from the more recent and better-written material on this page. QVVERTYVS (hm?) 11:12, 4 August 2013 (UTC)
There is a major problem of this article. It only covers the use in computer science. There are biological neural network that are artificially created. See here for an example: Implanted neurons, grown in the lab, take charge of brain circuitry.
Also, in computer science, the term, neural network, is very established. Major universities use NN instead of ANN as the name of subjects. Here is an example: http://www.cse.unsw.edu.au/~cs9444/ It should be renamed to neural network(computer).
My views of merge with other articles can be found on the talk page of neural network. Science.philosophy.arts (talk) 01:45, 20 September 2013 (UTC)
While looking at the article, I realized that the "Recent improvements" and the "Successes in pattern recognition contests since 2009" sections are very similar. For instance, a quote from the former section:
Such neural networks also were the first artificial pattern recognizers to achieve human-competitive or even superhuman performance[21] on benchmarks such as traffic sign recognition (IJCNN 2012), or the MNIST handwritten digits problem of Yann LeCun and colleagues at NYU.
And the latter:
Their neural networks also were the first artificial pattern recognizers to achieve human-competitive or even superhuman performance[21] on important benchmarks such as traffic sign recognition (IJCNN 2012), or the MNIST handwritten digits problem of Yann LeCun at NYU.
Wow. Since the former section is better integrated into the article and the latter section seems to be only something tacked on at the end, beginning with the slightly NPOVy phrase "[the] neural networks developed in the research group of Jürgen Schmidhuber at the Swiss AI Lab IDSIA have won eight international competitions", I would strongly recommend that the latter section be deleted and its content merged into the former section (this process seems to have been halfway carried out already). Comments? APerson (talk!) 02:20, 21 December 2013 (UTC)
"Also key later advances was the backpropagation algorithm which effectively solved the exclusive-or problem (Werbos 1975).[6]"
The Backpropagation algorithm doesn't solve the Xor problem, it allows efficient training of neural networks. It's just that a neural network can solve the Xor problem while a single neuron/perceptron can't.
217.140.110.23 (talk) 13:07, 15 April 2014 (UTC)Taylor
it's not clear to what degree artificial neural networks mirror brain function
I would take out this sentence as it is 100% clear brain doesn't compute gradients. Mosicr (talk) 16:10, 13 September 2016 (UTC)
Artificial neural network has various application in production or manufacturing[1] that are capable of machine learning[2] & pattern recognition[3]. Various machining[4] processes require prediction of various results on the basis of the input data or quality[5] characteristics[6] provided in the machining process & similarly back tracking of required quality characteristics for a given result or desired output characteristics.--Rahulpratapsingh06 (talk) 12:39, 5 May 2014 (UTC)
References
{{cite web}}
: Missing or empty |title=
(help)
{{cite web}}
: Missing or empty |title=
(help)
{{cite web}}
: Missing or empty |title=
(help)
{{cite web}}
: Missing or empty |title=
(help)
{{cite web}}
: Missing or empty |title=
(help)
{{cite web}}
: Missing or empty |title=
(help)
The relationship between various quality characteristics & outputs can be learned by the artificial neural network design on the basis of the algorithms and programing over the data provided, which is machine learning or pattern recognition.--Rahulpratapsingh06 (talk) 12:29, 5 May 2014 (UTC)
I remove this part : "Some may be as simple as a one-neuron layer with an input and an output, and others can mimic complex systems such as dANN, which can mimic chromosomal DNA through sizes at the cellular level, into artificial organisms and simulate reproduction, mutation and population sizes.[1]" because dANN is not popular. What do you think ? --Vinchaud20 (talk) 10:05, 19 May 2014 (UTC)
Also " Artificial neural networks can be autonomous and learn by input from outside "teachers" or even self-teaching from written-in rules." should be remove because it is a reformulation of the learning process. And here, we speak about the "Type of Neural network" and not the "learning process" --Vinchaud20 (talk) 10:12, 19 May 2014 (UTC).
References
{{cite web}}
: Unknown parameter |deadurl=
ignored (|url-status=
suggested) (help)
I want to know about fluidization
recent improvements and successes since 2009 are nearly identical. I think the since 2009 section is obsolete
LuxMaryn (talk) 13:26, 26 November 2014 (UTC)
I am trying to imagine someone like a very bright high school student who heard that neural networks might be interesting, and visited this article to learn at least a little about them.
The student will learn nothing whatsoever about neural networks from this article. They will learn, however, that many people who write for Wikipedia have not the slighest idea of what an encyclopedia article ought to be like.
The text does not explain anything to anyone who doesn't already know what neural nets are. There is not even one — not even one — example of a simple neural net for someone who has never seen one before. All the inscrutable definitions and diagrams do nothing at all toward helping a newcomer to the subject understand what a neural net is.Daqu (talk) 01:40, 8 May 2015 (UTC)
I know a lot of math and statistics, but I honestly still don't understand the concept of a neural network after looking at this article. Looking at this talk page, it's obvious that a lot of people are dissatisfied with it. I think that, as suggested by the original poster in this thread, a lot of improvement could be made simply by putting in, immediately after the lead, a very simple example of a neural network, with all details included—each variable defined explicitly, each neuron defined explicitly, etc. Loraof (talk) 20:34, 14 September 2015 (UTC)
Neural networks are so hot right now, there are so many incredibly exciting applications out there, and this article barely mentions one. Examples, we need examples. Stupid practical examples to depict how a simple NN works, and interesting examples of possible applications to show what it is possible to achieve with NN (self-piloting helicopter anyone?) 193.205.78.232 (talk) 16:47, 7 October 2015 (UTC)
The wikipedia article states that neural networks and directed graphical models are ``largely equivalent``. While I know both feed forward neural networks and directed graphical models, I don't understand where this equivalence should come from (admittedly both models seem to be similar enough to suspect something like this though). Could anyone elaborate on this or at least add a source with the precise meaning of this statement? — Preceding unsigned comment added by 77.9.130.141 (talk) 07:03, 6 November 2015 (UTC)
See Reverse engineering. Example. Modeling of the nervous system of reptiles (Russian) (and kangaroo): F-22 Raptor Main article is bad. Much math without physical meaning. RippleSax (talk) 16:00, 1 December 2015 (UTC)
The article cites Hebbian learning as one of the origins of Artificial Neural Networks. The only related citation in this article, the Wikipedia entry for Hebbian learning, and my own research indicate that Hebb's oldest work on this topic was published in 1949.
Simultaneously, the article states that "Researchers started applying [Hebbian learning] to computational models in 1948 with Turing's B-type machines."
Neither Hebbian learning prior to 1949 or the 1948 models are cited and it seems to imply that the ideas published by Hebb in 1949 were already being applied a year earlier in 1948. Thedatascientist (talk) 16:59, 20 January 2016 (UTC)
The "Theoretical" section of the article is badly in need of revision. Specifically, this excerpt: "Nothing can be said in general about convergence since it depends on a number of factors. Firstly, there may exist many local minima. This depends on the cost function and the model. Secondly, the optimization method used might not be guaranteed to converge when far away from a local minimum. Thirdly, for a very large amount of data or parameters, some methods become impractical. In general, it has been found that theoretical guarantees regarding convergence are an unreliable guide to practical application."
* "There may exist many local minima." How does this affect convergence? * "The optimization method used might not be guaranteed to converge when far away from a local minimum." Example? * "Some methods become impractical." Which ones? * "In general, it has been found that theoretical guarantees regarding convergence are an unreliable guide to practical application." By who?
97.79.173.131 (talk) 01:44, 7 June 2016 (UTC)j_Kay
Dr. Gallo has reviewed this Wikipedia page, and provided us with the following comments to improve its quality:
This article is well organized and written, with an adequate level of detail. No inaccuracies or errors seem to be present.
We hope Wikipedians on this talk page can take advantage of these comments and improve the quality of the article accordingly.
Dr. Gallo has published scholarly research which seems to be relevant to this Wikipedia article:
ExpertIdeasBot (talk) 13:41, 11 June 2016 (UTC)
Maybe these are buried somewhere in the article, but I can't figure them out. I think a new section should be added near the beginning to address these questions.
Loraof (talk) 17:08, 17 July 2016 (UTC)
When I read this article just now, the article referred (in History-->Improvements since 2006) to the simple and complex cells in the visual cortex discovered by David Hubel and L. Ron Hubbard. I'm not an expert in the field, but as far as I can tell, Hubel's partner in that research was Torsten Wiesel. I can't find any reliable source that mentions Hubbard studying neurology or vision, and I suspect that it was either a mistake or vandalism. I've corrected it to Torsten Wiesel. — Preceding unsigned comment added by 91.135.102.171 (talk) 15:39, 26 September 2016 (UTC)
I agree with all the comments that this is a lot of detailed information without a good overview. I will work on an introduction to it all that makes it a bit more clear. — Preceding unsigned comment added by Anximander (talk • contribs) 08:57, 16 October 2016 (UTC)
I can see a number of similarities between cellular automata and neural networks. Is there a known relationship between these two models? Are they computationally equivalent/inequivalent? 75.139.254.117 (talk) 13:47, 18 March 2017 (UTC)
I have a degree in mathematics. I am understanding the math parts of this just fine, as I think anyone with a general grasp of college level mathematics will. However, I have found it very difficult to understand how the math set forth in this article corresponds to what is actually happening when an ANN-modeled computer is trying to compute something. Mainly because this article doesn't explain what activation or inhibition mean mathematically, or even what they are conceptually. I think. Are the neurons in the input layer observing individual elements of a vector, or are they observing the whole vector but competing with each other because they are taking slightly different values? What's going on in the "hidden" layers? Are the output layers putting out individual elements of a vector, or something else? I'm not even sure if these questions make sense. 100.33.30.170 (talk) 16:06, 5 April 2017 (UTC)
Update: I tagged the models section to reflect this and added some explain tags to some sentences. 100.33.30.170 (talk) 16:17, 5 April 2017 (UTC)
Update: What would help most if there was a very basic practical example of a neural network computing something. 100.33.30.170 (talk) 16:20, 5 April 2017 (UTC)
Very helpful in improving the clarity of what the neural network is actually computing. However the next section seems to be referring to similar mathematical functions using different notation. For instance in components section, the activation function is written as a_j(t), whereas in the next section the activation function is referred to as K. Perhaps this article should be edited to unify the notation as I think this creates an unnecessary confusion. Or if they are different then explain why they are different. 38.140.162.114 (talk) 16:15, 3 August 2017 (UTC)
The Neural Computation redirect sends here, but the concept is well established in its own right. It is more closely related to the page 'Models of Neural Computation' or 'Biological Neural Networks'. Artificial Neural Networks as in this present article are but a subset of neural computation which took flight (through its applications in machine learning).
I propose editing the disambiguation page for Neural Computation and splitting the concept between this page, the eponymous journal, or 'Models of Neural Computation'. The redirect eclipses the good article on 'Models of Neural Computation', of more interest to neuroscientists.
MNegrello (talk) 12:26, 28 August 2017 (UTC)
A large amount of edited templates were inserted that essentially put talk page opinions into the body of the article. These opinions should be moved to the talk page.
I think that the comments look like something I myself would write.....in essence that the text does not do a good job of teaching the topic. However, when writing such things I've received responses that Wikipedia does not teach, it provides information. I disagree with a categorical interpretation of that. So I agree with the comments (once moved to the talk page), they are iffy in Wikipedia.
Finally, persons with such thoughts should work on the article.
Sincerely, North8000 (talk) 23:38, 23 October 2017 (UTC)
I read this paragraph 3 times and did not find anything of meaning or merit in it. So I made an edit removing it. Please re-add it if you will take on the task of actually making it say something meaningful and well cited.
Sincerely Patriotmouse (talk)
User:Nbro recently added a few dozen cleanup tags to this article, but they appear to be mostly redundant. Would it be possible to rollback these redundant tags instead of removing them manually? Jarble (talk) 18:58, 15 November 2017 (UTC)
As the article Pattern recognition itself suggests, arguably any type of machine learning can be understood as pattern recognition, so I don't know what the category "pattern recognition (radar systems, face identification, signal classification,[203] object recognition and more)" in the list of applications is supposed to tell me. — Preceding unsigned comment added by 141.84.69.83 (talk) 18:29, 11 July 2018 (UTC)
1) Basically, the recent papers which affect the advances n neural nets should be mentioned. Specifically papers like properties of neural nets which discusses how neural nets can be fooled by just adding noise so that any image could be classified as ostrich.
2) New architectures like attention based models and new areas like forensic where it is applied is not discussed in the article. — Preceding unsigned comment added by Shubhamag97 (talk • contribs) 03:27, 13 September 2018 (UTC)
I have dumped two sizable sections from backpropagation in this article. I am not sure that this improved this article. I am sure though, that it improved the Wikipedia, because the texts became neither worse nor better and the material does belong in the topic of this article. I know the problem of dividing material up between ANN and deep learning is severe. But again, learning is not part of BP. Maybe we should have a separate article about artificial neural network learning. --Ettrig (talk) 13:59, 28 November 2018 (UTC)
Someone has dumped a small paragraph about capsule networks into the section about convolutional networks, "A recent development has been that of Capsule Neural Network (CapsNet)…"[1] This is not quite right as the capsule network in Hintons variant has a routing component, and this does not fit very well convolutional networks. The convoluting action in the first capsule layer is somewhat similar to a convolutional network, but the routing has more in common with recurrent networks. There are other differences too, but the previous is perhaps the simplest way to see how they diverge.
An alternative way to describe the differences is to say a convolutional network is a stateless network, the transform from the input to the output is a stateless function, while a capsule network manage an internal state. It needs the internal state to converge to a solution.
I would say Hintons capsule network is a specific type from a new class of networks, where the class is "networks that correlates normalized data". The normalization can be done in several ways over several layers. It can even be learned in deep networks.
[Addition: A fun thing is that Hintons capsules does the correlation in the routing, and nearly avoids the normalization, but by doing so it perpetuates the core problem, it must learn the new overall pose. It seems like the cortex avoids this problem altogether and learn a general pose.] Jeblad (talk) 13:18, 31 March 2019 (UTC)
I noticed that the first three papers in the section on Memory networks are cited to be authored by Jürgen Schmidhuber, while in the papers he does not appear as an author.
Meerpirat (talk) 09:36, 4 April 2019 (UTC)
[2] — Preceding unsigned comment added by 14.177.144.110 (talk) 00:09, 3 June 2019 (UTC)
Hi @Tryptofish, I added a sentence at the top of Variants section, and linked to the neural network zoo. Why did you revert my edit? Your note mentions WP:COI, but I do not know the zoo authors, and I am not affiliated with them. The catalogue that they have created helps the reader have a better understanding of different ANN architectures. --Habil zare (talk) 21:56, 26 June 2019 (UTC)
I am going to stay on topic and simply state the reason why, in my view, this page will never be acceptable. in other words, I'll discuss improvements to the page and nothing else.
Since there is no such thing as an artificial neural network (a lot of stuff related to Artificial Intelligence is pure bull), no one can explain it in layman's terms, in simple terms and all you're going to get is difficult explanations that don't make any sense. Such as this one: https://arxiv.org/pdf/1609.08144.pdf?fbclid=IwAR21rxrFrNqJ3G-flYcqbpUbhG79ChD9DBG8uzo9htlnu-dXhAWaaKwBuGw
AI people will pretend that their concept is so complex that only super brains can understand it, which is bull of course. There is a lot of bull concerning artificial intelligence. Why? Researchers want to keep money pouring in in research projects so they pretend they have accomplished a lot, which is a lie. In my view, at one point, Governments will step in and investigate for a lot of actions perpetrated by AI people are close to fraud and thousand of investors are investing into a field that hypes itself and lives on false claims.
So in the end, you'll have the following choice:
• a) Publish a complex explanation that does not make sense
• b) Publish nothing for you won't find a simple explanation anywhere.
Since you don't want a) there is only b) left. You have no other choice.
Therefore the best thing to do in my view, is to simply state in the article that since there is no reference available that explains the concept in layman's terms, you refrained to provide an explanation coming from the Artificial intelligence milieu because no one could understand it.
If you're expecting a simple and plausible explanation to insert in the article, an explanation you can reference, don't hold your breadth because it just won't happen in a million years. So forget it!
-- Robert Abitbol — Preceding unsigned comment added by 24.54.3.100 (talk) 18:20, 15 December 2019 (UTC)
-- Robert Abitbol
— Preceding unsigned comment added by 24.54.3.100 (talk) 19:42, 19 December 2019 (UTC)
This change changes two things. One of them has a motivation that is clearly wrong. That can be seen by a look at the detailed article about the history. That article also shows that the revert introduces an error. The other change back has a motivation that does not respond to the motivatione for the change (in the change comment) that it reverts.--Ettrig (talk) 15:08, 18 May 2020 (UTC)