Skip to main content

Neural Networks

Neural Networks are also known as Artificial Neural Networks (ANN). They are at the heart of deep learning algorithms, and are inspired by the human brain, trying to replicate how neurons signal each other.

In RL, Neural Networks are used with a variety of algorithms, like Q-Learning for example. v^\hat{v} can be a function computed by a multi-layer ANN, with w\bold{w} the vector of connection weights in all the layers.

ANNs are comprised of node layers.

  • An input layer
  • One or more hidden layers
  • An output layer

Each node (or artificial neuron) connects to another (in the neighbouring layer) and has an associated weight and threshold. If the output of any individual node is above the threshold value, that node is activated, sending the data to the next layer of the network. Otherwise, no data is passed along.

y=โˆ‘i=1mwixi+bias\large y = \sum_{i=1}^{m} w_i x_i + \textsf{bias}

The Activation Function which we generally use is called ReLU (Rectified Linear Unit)

ReLU(x)=maxโก(0,x)\textsf{ReLU}(x) = \max(0, x)

Finally, we define some kind of Cost Function based on MSE and do SGD to converge to a solution (for weights).

Reference: https://www.ibm.com/in-en/cloud/learn/neural-networks