Neural Nets 2: Difference between revisions
Bradley Monk (talk | contribs) No edit summary |
Bradley Monk (talk | contribs) No edit summary |
||
Line 7: | Line 7: | ||
4. [[Neural Nets 4|Network Outputs]]<br> | 4. [[Neural Nets 4|Network Outputs]]<br> | ||
}} | }} | ||
<br><br><br><br><br><br> | |||
<br><br><br><br><br><br> | |||
==TBD BELOW== | |||
Below I've embedded another Tensorflow neural net playground. | Below I've embedded another Tensorflow neural net playground. |
Revision as of 13:46, 22 January 2018
(Back to Tensorflow Tutorial Page 1)
TUTORIAL ON MACHINE LEARNING AND NEURAL NETWORKS (PAGE 2)
TBD BELOW
Below I've embedded another Tensorflow neural net playground.
{{#widget:Tensorflow}}
Outputs
More directly, it is the value spit-out by the activation function of the 'output layer'. Here, since we only have a single layer, our hidden 'hidden layer' and 'output layer' are one in the same. The output function of our neuron is known as the tanh function.
The tanh function is an extremely common choice for an output function in artificial neural network machine learning frameworks because it yields a nice sigmoid shape, and no matter the magnitude of its inputs, the output from the tanh function is bounded between { 0 : 1}. These are very desirable properties for neural net nodes. Here you see the tanh function evaluated across various x-dim inputs...
Tanh produces a sigmoid output over the range {-2 : 2}, and automatically evaluates to exact values when its argument is the natural logarithm. Speaking of the natural log, that is another very common choice of output function for the same reasons as tanh.
For now, let's not belabor the point that our neuron (and in going forward, all our neurons) are using the tanh function. Maybe just keep this in mind if you're wondering what sorts of numbers are travelling along the axons of these neurons, and ultimately those colored gradients underneath the dots.
This tutorial continues on the next page. Don't worry about playing around too much with the TensorFlow GUI, there will be plenty of that on the next page, and those that follow.