-
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Hi @avish121 , Excellent questions!
This sounds plausible. I'd agree too. With an infinite amount of straight lines at almost any angle/slope you should be able to draw any pattern you please. In fact, this is how I believe the Linear SVM (another machine learning function) works, by separating data linearly across different planes. However, in practice, neural networks function far better by combining linear and non-linear functions. As to exactly why, I don't have a perfectly good answer (I've searched for them myself), other than the explanation you linked (combine straight and non-straight lines). It is only through experience of building 1000s of models have I (and many others) come to a similar conclusion. There may be a mathematical proof somewhere but I am unaware of it (if you find it, I'd love to read it).
Yes that's correct (though sometimes not all neurons are connected to the next layer, such as in the case of dropout). Each neuron can be considered it's own function (e.g. Combine the learned patterns of each individual neuron and hopefully you get a good representation of the data you're working with. As for your question about how patterns in the first layer are related to the final output. Generally, each layer progressively learns more refined patterns. Starting from the rougher straight lines in the first layer (e.g. 3x straight lines in your example trying to cut up the data). To the more refined circles in the second layer. This kind of example can be seen here because our target data is quite easily visually separable (e.g. you can see with your eyes the blue dots should be separate from the orange dots). However, in larger datasets, this kind of visualization is often not possible (because there are far too many dimensions). And generally, with larger datasets, you use larger neural networks with more individual neurons. So inspecting what an individual neuron learns on a large dataset becomes harder and harder with size. However, to see a cool example of an individual neuron learning something, you may be interested in reading OpenAI's paper called "Unsupervised Sentiment Neuron", where it was found out of thousands of individual neurons (4096 total), one was found to learn the sentiment of a piece of text. Anyway, fantastic questions and let me know if you'd to discuss more or me to expand on anything. Daniel |
Beta Was this translation helpful? Give feedback.
Hi @avish121 ,
Excellent questions!
This sounds plausible. I'd agree too.
With an infinite amount of straight lines at almost any angle/slope you should be able to draw any pattern you please.
In fact, this is how I believe the Linear SVM (another machine learning function) works, by separating data linearly across different planes.
However, in practice, neural networks function far better by combining linear and non-linear functions.
As to exactly why, I don't …