Deep Neural Networks are nonlinear Activation Functions

Almost any process imaginable can be represented as a functional computation in a neural network, provided that the activation function is non-linear.

Non-linear functions address the problems of a linear activation function:

  1. They allow backpropagation because they have a derivative function which is related to the inputs.

  2. They allow “stacking” of multiple layers of neurons to create a deep neural network. Multiple hidden layers of neurons are needed to learn complex data sets with high levels of accuracy.

Recent Posts

See All

ML Algorithms Comparative Analysis Overview Classification Algorithms Logistic Regression Naive Bayes Classifier Regression Algorithms Linear Regression Either Classification or Regression Algorithms

A drug target is a molecule in the body, usually a protein, that is intrinsically associated with a particular disease process and that could be addressed. A biological target is anything within a liv