## Overview

- Neural Networks are one of the most popular machine learning algorithms
- Gradient Descent forms the basis of Neural networks
- Neural networks can be implemented in both R and Python using certain libraries and packages

## Introduction

You can learn and practice a concept in two ways:

**Option 1:**You can learn the entire theory on a particular subject and then look for ways to apply those concepts. So, you read up how an entire algorithm works, the maths behind it, its assumptions, limitations and then you apply it. Robust but time taking approach.**Option 2:**Â Start with simple basics and develop an intuition on the subject. Next, pick a problem and start solving it. Learn the concepts while you are solving the problem. Keep tweaking and improving your understanding. So, you read up how to apply an algorithm – go out and apply it. Once you know how to apply it, try it around with different parameters, values, limits and develop an understanding of the algorithm.

I prefer Option 2 and take that approach to learning any new topic. I might not be able to tell you the entire math behind an algorithm, but I can tell you the intuition. I can tell you the best scenariosÂ to apply anÂ algorithm based on my experiments and understanding.

In my interactions with people, I find that people don’t take time to develop this intuition and hence they struggle to apply things in the right manner.

In this article, I will discuss the building block of a neural network from scratch and focus more on developingÂ this intuition to apply Neural networks. WeÂ will code in both “Python” and “R”. By end of this article, you will understand how Neural networks work, how do we initialize weigths and how do we update them usingÂ back-propagation.

Let’s start.

## Table of Contents:

- Simple intuition behind Neural networks
- Multi Layer Perceptron and its basics
- Steps involved in Neural Network methodology
- Visualizing steps for Neural Network working methodology
- Implementing NN using Numpy (Python)
- Implementing NN using R
- [Optional] Mathematical Perspective of Back Propagation Algorithm

## Simple intuition behind neural networks

If you have been a developer or seen one work – you know how it is to search for bugs in a code. You would fire various test cases by varying the inputs or circumstances and look for the output. The change in output provides you a hint on where to look for the bug – which module to check, which lines to read. Once you find it, you make the changes and the exercise continues until you have the right code / application.

Neural networks work in very similar manner.Â It takes several input, processes it through multiple neurons from multiple hidden layers and returns the result using an output layer. This result estimation process is technically known as “**Forward Propagation**“.

Next, we compare the result with actual output. The task is to make the output to neural network as close to actual (desired) output. Each ofÂ these neurons are contributing some error to final output. How do you reduce the error?

We try to minimize the value/ weight of neurons those are contributing more to the error and this happens while traveling back to the neurons of the neural network and finding where the error lies. This process is known as “**Backward Propagation**“.

In order to reduce these number of iterations to minimize the error, the neural networksÂ use a common algorithm known as “Gradient Descent”, which helps to optimize the task quickly and efficiently.

That’s it – this is how Neural network works! I know this is a very simple representation, but it would help you understand things in a simple manner.

## Multi Layer Perceptron and its basics

Just like atoms form the basics of any material on earth – the basic forming unit of a neural network is a perceptron. So, what is a perceptron?

A perceptron can be understood as anything that takes multiple inputs and produces one output. For example, look at the image below.

The above structure takes three inputs and produces one output. The next logical question is what is the relationship between input and output? Let us start with basic ways and build on to find more complex ways.

Below, I have discussed three ways of creating input output relationships:

**By directly combining the input and computing the output**based on a threshold value. for eg: Take x1=0, x2=1, x3=1 and setting a threshold =0. So, if x1+x2+x3>0, the output is 1 otherwise 0. You can see that in this case, the perceptron calculates the output as 1.**Next, let us add weights to the inputs.**Â Weights give importance to an input. For example, you assign w1=2, w2=3 and w3=4 to x1, x2 and x3 respectively. To compute the output, we will multiply input with respective weights and compare with thresholdÂ value as w1*x1 + w2*x2 + w3*x3 > threshold. These weights assign more importance to x3 in comparison to x1 and x2.Each perceptronÂ also has a bias which can be thought of as how much flexible the perceptronÂ is. It is somehow similar to the constant**Next, let us add bias:***b*of a linear functionÂ*y = ax + b. It allows us to move the line up and down to fit the prediction with the data better. Without*Â For example, a perceptronÂ may have two inputs, in that case, it requires three weights. One for each input and one for the bias. Now linear representation of input will look like,Â w1*x1 + w2*x2 + w3*x3 + 1*b.*b*the line will always goes through the origin (0, 0) and you may get a poorer fit.

But, all of this is still linear which is what perceptrons used to be. But that was not as much fun. So, people thought of evolving a perceptron to what is now called as artificial neuron. A neuron applies non-linear transformations (activation function) to the inputs and biases.

### What is an activation function?

Activation FunctionÂ takes the sum of weighted input (w1*x1 + w2*x2 + w3*x3 + 1*b) as an argument and return the output of the neuron.Â In above equation, we have represented 1 as x0 and b as w0.

The activation function is mostly used to make a non-linear transformation which allows us to fit nonlinear hypotheses or to estimate the complex functions. There are multiple activation functions, like: “Sigmoid”, “Tanh”, ReLu and many other.

### Forward Propagation, Back Propagation and Epochs

Till now, we have computed the output and this process is known as “**Forward Propagation**“. But what if the estimated output is far away from the actual output (high error). In the neural network what we do, we update the biases and weights based on the error. This weight and bias updating process is known as “**Back Propagation**“.

Back-propagation (BP) algorithms work by determining the loss (or error) at the output and then propagatingÂ it back into the network. The weights are updated to minimize the error resulting from each neuron. The first step in minimizing the error is to determine the gradient (Derivatives) of each node w.r.t. the final output. To get a mathematical perspective of the Backward propagation, refer below section.

This one round of forward and back propagation iteration is known as one training iteration aka “**Epoch**“.

### Multi-layer perceptron

Now, let’s move on to next part of **Multi-Layer** Perceptron. So far, we have seen just a single layer consisting of 3 input nodes i.e x1, x2 and x3 and an output layer consisting of a single neuron. But, for practical purposes, the single-layer network can do only so much. An MLP consists of multiple layers called **Hidden LayersÂ **stacked in between the **Input Layer** and the **Output Layer** as shown below.

The image above shows just a single hidden layer in green but in practice can contain multiple hidden layers. Another point to remember in case of an MLP is that all the layers are fully connected i.e every node in a layer(except the input and the output layer) is connected to every node in the previous layer and the following layer.

Let’s move on to the next topic which is training algorithm for a neural network (to minimize the error). Here, we will look at most common training algorithm known asÂ Gradient descent.

### Full Batch Gradient Descent and Stochastic Gradient Descent

Both variants of Gradient Descent perform the same work of updating the weights of the MLP by using the same updating algorithm but the difference lies in the number of training samples used to update the weights and biases.

Full Batch Gradient Descent Algorithm as the name implies uses all the training data points to update each of the weights once whereas Stochastic Gradient uses 1 or more(sample) but never the entire training data to update the weights once.

Let us understand this with a simple example of a dataset of 10 data pointsÂ with two weights **w1** and **w2**.

**Full Batch:** You use 10 data points (entire training data) and calculate the change in w1 (Î”w1) and change in w2(Î”w2) and update w1 and w2.

**SGD:** You use 1st data point and calculate the change in w1 (Î”w1) and change in w2(Î”w2) and update w1 and w2. Next, when you use 2nd data point, you will work on the updated weights

For a more in-depth explanation of both the methods, you can have a look atÂ this article.

## StepsÂ involved in Neural Network methodology

Let’s look at the step by step building methodology of Neural Network (MLP with one hidden layer, similar to above-shown architecture). At the output layer, we have only one neuron as we are solving a binary classification problem (predict 0 or 1). We could alsoÂ have two neurons for predicting each of both classes.

First look at the broad steps:

0.) We take input and output

- X as an input matrix
- y as an output matrix

1.) We initialize weights and biases with random values (This is one time initiation. In the next iteration, we will use updated weights, and biases). Let us define:

- wh as weightÂ matrix to the hidden layer
- bh as bias matrix to the hidden layer
- wout asÂ weightÂ matrix to the output layer
- bout as bias matrix to the output layer

2.) We take matrix dot product of input and weights assigned to edges between the input and hidden layer then add biases of the hidden layer neurons to respective inputs, this is known as linearÂ transformation:

hidden_layer_input= matrix_dot_product(X,wh) + bh

3) Perform non-linear transformation using an activation function (Sigmoid). Sigmoid will return the output asÂ 1/(1 + exp(-x)).

hiddenlayer_activations = sigmoid(hidden_layer_input)

4.) Perform a linear transformation on hidden layer activation (take matrix dot product with weights and add a bias of the output layer neuron) then apply an activation function (again used sigmoid, but you can use any other activation function depending upon your task) to predict the output

output_layer_input = matrix_dot_product (hiddenlayer_activations * wout ) + bout

*output = sigmoid(output_layer_input)
*

**All above steps are known as “Forward Propagation“**

5.) Compare prediction with actual output and calculate the gradient of error (Actual – Predicted). Error is the mean square loss = ((Y-t)^2)/2

E = y – output

6.) Compute the slope/ gradient of hidden and output layer neurons (Â To compute the slope, we calculate the derivatives of non-linear activations x at each layer for each neuron). Gradient of sigmoid can be returned asÂ x * (1 – x).

slope_output_layer = derivatives_sigmoid(output)

*slope_hidden_layer = derivatives_sigmoid(hiddenlayer_activations)*

7.) ComputeÂ change factor(delta) at output layer, dependent on the gradient of error multiplied by the slope of output layer activation

*d_output = E * slope_output_layer*

8.) At this step, the error will propagate back into the network which means error at hidden layer. For this, we will take the dot product of output layer delta with weight parameters of edges between the hidden and output layer (wout.T).

Error_at_hidden_layer = matrix_dot_product(d_output, wout.Transpose)

9.) ComputeÂ change factor(delta) at hidden layer, multiply the error at hidden layer with slope of hidden layer activation

*d_hiddenlayer = Error_at_hidden_layer * slope_hidden_layer*

10.) Update weights at the output and hidden layer: The weights in the network can be updated from the errors calculated for training example(s).

wout = wout + matrix_dot_product(hiddenlayer_activations.Transpose, d_output)*learning_rate

wh = Â wh + matrix_dot_product(X.Transpose,d_hiddenlayer)*learning_rate

learning_rate: The amount that weights are updated is controlled by a configuration parameter called the learning rate)

11.) Update biases at the output and hidden layer: The biases in the network can be updated from the aggregated errors at that neuron.

- bias at output_layer =bias at output_layer + sum of deltaÂ of output_layer at row-wise * learning_rate
- bias at hidden_layer =bias at hidden_layer +Â sum of deltaÂ of output_layer at row-wise * learning_rate
*Â Â*

*bh = bh + sum(d_hiddenlayer, axis=0) * learning_rate
*

*bout = bout + sum(d_output, axis=0)*learning_rate*

**Steps from 5 to 11 are known as “Backward Propagation“**

One forward and backward propagation iteration is considered as one training cycle. As I mentioned earlier, When do we train second time then update weights and biases are used for forwardÂ propagation.

Above, we have updated the weight and biases for hidden and output layer and we have used full batch gradient descent algorithm.

## Visualization of steps for Neural Network methodology

We will repeat the above steps and visualize the input, weights, biases, output, error matrix to understand working methodology of Neural Network (MLP).

Note:

- For good visualization images, I have rounded decimal positions at 2 or3 positions.
- Yellow filled cells represent current active cell
- Orange cell represents the input used to populate values of current cell

**Step 0: **Read input and output

**Step 1: **Initialize weights and biases with random values (There are methods to initialize weights and biases but for now initialize with random values)

**Step 2:** Calculate hidden layer input:

hidden_layer_input= matrix_dot_product(X,wh) + bh

**Step 3:Â **Perform non-linear transformation on hidden linear input

*hiddenlayer_activations = sigmoid(hidden_layer_input)*

**Step 4:Â **Perform linear and non-linear transformation of hidden layer activation at output layer

output_layer_input = matrix_dot_product (hiddenlayer_activations * wout ) + bout

*output = sigmoid(output_layer_input)*

**Step 5: **Calculate gradient of Error(E) at output layer

*E = y-output*

**Step 6:Â **Compute slope at output and hidden layer

*Slope_output_layer= derivatives_sigmoid(output)*

*Slope_hidden_layer = derivatives_sigmoid(hiddenlayer_activations)*

**Step 7: **Compute delta at output layer

*d_output = E * slope_output_layer*lr*

**Step 8: **Calculate Error at hidden layer

Error_at_hidden_layer = matrix_dot_product(d_output, wout.Transpose)

**Step 9:Â **Compute delta at hidden layer

*d_hiddenlayer = Error_at_hidden_layer * slope_hidden_layer*

**Step 10: **Update weight at both output and hidden layer

wout = wout + matrix_dot_product(hiddenlayer_activations.Transpose, d_output)*learning_rate

wh = Â wh+ matrix_dot_product(X.Transpose,d_hiddenlayer)*learning_rate

**Step 11: **Update biases at both output and hidden layer

*bh = bh + sum(d_hiddenlayer, axis=0) * learning_rate
*

*bout = bout + sum(d_output, axis=0)*learning_rate*

Above, you can see that there is still a good error not close to actual target value because we have completed only one training iteration. If we will train model multiple times then it will be a very close actual outcome. I have completed thousands iteration and my result is close to actual target values ([[ 0.98032096] [ 0.96845624] [ 0.04532167]]).

## Implementing NN using Numpy (Python)

## Implementing NN in R

`# input matrix`

`X=matrix(c(1,0,1,0,1,0,1,1,0,1,0,1),nrow = 3, ncol=4,byrow = TRUE)`

`# output matrix`

`Y=matrix(c(1,1,0),byrow=FALSE)`

`#sigmoid function`

`sigmoid<-function(x){`

`1/(1+exp(-x))`

`}`

`# derivative of sigmoid function`

`derivatives_sigmoid<-function(x){`

`x*(1-x)`

`}`

`# variable initialization`

`epoch=5000`

`lr=0.1`

`inputlayer_neurons=ncol(X)`

`hiddenlayer_neurons=3`

`output_neurons=1`

`#weight and bias initialization`

`wh=matrix( rnorm(inputlayer_neurons*hiddenlayer_neurons,mean=0,sd=1), inputlayer_neurons, hiddenlayer_neurons)`

`bias_in=runif(hiddenlayer_neurons)`

`bias_in_temp=rep(bias_in, nrow(X))`

`bh=matrix(bias_in_temp, nrow = nrow(X), byrow = FALSE)`

`wout=matrix( rnorm(hiddenlayer_neurons*output_neurons,mean=0,sd=1), hiddenlayer_neurons, output_neurons)`

`bias_out=runif(output_neurons)`

`bias_out_temp=rep(bias_out,nrow(X))`

`bout=matrix(bias_out_temp,nrow = nrow(X),byrow = FALSE)`

`# forward propagation`

`for(i in 1:epoch){`

`hidden_layer_input1= X%*%wh`

`hidden_layer_input=hidden_layer_input1+bh`

`hidden_layer_activations=sigmoid(hidden_layer_input)`

`output_layer_input1=hidden_layer_activations%*%wout`

`output_layer_input=output_layer_input1+bout`

`output= sigmoid(output_layer_input)`

`# Back Propagation`

`E=Y-output`

`slope_output_layer=derivatives_sigmoid(output)`

`slope_hidden_layer=derivatives_sigmoid(hidden_layer_activations)`

`d_output=E*slope_output_layer`

`Error_at_hidden_layer=d_output%*%t(wout)`

`d_hiddenlayer=Error_at_hidden_layer*slope_hidden_layer`

`wout= wout + (t(hidden_layer_activations)%*%d_output)*lr`

`bout= bout+rowSums(d_output)*lr`

`wh = wh +(t(X)%*%d_hiddenlayer)*lr`

`bh = bh + rowSums(d_hiddenlayer)*lr`

`}`

`output`

## [Optional] Mathematical Perspective of Back Propagation Algorithm

Let W_{i }be the weights between the input layer and the hidden layer. W_{h }be the weights between the hidden layer and the output layer.

Now, **h=Ïƒ (u)= Ïƒ (W _{i}X)**, i.e h is a function of u and u is a function of W

_{i }and X. here we represent our function as

**Ïƒ**

**Y= Ïƒ (uâ€™)= Ïƒ (**W** _{h}**h

**)**, i.e Y is a function of uâ€™ and uâ€™ is a function of W

_{h }and h.

We will be constantly referencing the above equations to calculate partial derivatives.

We are primarily interested in finding two terms, âˆ‚E/âˆ‚W_{i }and âˆ‚E/âˆ‚W_{h }i.e change in Error on changing the weights between the input and the hidden layer and change in error on changing the weights between the hidden layer and the output layer.

But to calculate both these partial derivatives, we will need to use the chain rule of partial differentiation since E is a function of Y and Y is a function of uâ€™ and uâ€™ is a function of W_{i.}

Letâ€™s put this property to good use and calculate the gradients.

âˆ‚E/âˆ‚W_{h} = (âˆ‚E/âˆ‚Y).( âˆ‚Y/âˆ‚uâ€™).( âˆ‚uâ€™/âˆ‚W_{h}), â€¦â€¦..(1)

We know E is of the form E=(Y-t)^{2}/2.

So, (âˆ‚E/âˆ‚Y)= (Y-t)

Now, Ïƒ is a sigmoid function and has an interesting differentiation of the form Ïƒ(1- Ïƒ). I urge the readers to work this out on their side for verification.

So, (âˆ‚Y/âˆ‚uâ€™)= âˆ‚( Ïƒ(uâ€™)/ âˆ‚uâ€™= Ïƒ(uâ€™)(1- Ïƒ(uâ€™)).

But, Ïƒ(uâ€™)=Y, So,

(âˆ‚Y/âˆ‚uâ€™)=Y(1-Y)

Now, ( âˆ‚uâ€™/âˆ‚W_{h})= âˆ‚( W_{h}h)/ âˆ‚W_{h} = h

Replacing the values in equation (1) we get,

**âˆ‚E/âˆ‚W _{h} = (Y-t). Y(1-Y).h**

So, now we have computed the gradient between the hidden layer and the ouput layer. It is time we calculate the gradient between the input layer and the hidden layer.

âˆ‚E/âˆ‚W_{i} =(âˆ‚ E/âˆ‚ h). (âˆ‚h/âˆ‚u).( âˆ‚u/âˆ‚W_{i})

But, (âˆ‚ E/âˆ‚ h) = (âˆ‚E/âˆ‚Y).( âˆ‚Y/âˆ‚uâ€™).( âˆ‚uâ€™/âˆ‚h). Replacing this value in the above equation we get,

âˆ‚E/âˆ‚W_{i} =[(âˆ‚E/âˆ‚Y).( âˆ‚Y/âˆ‚uâ€™).( âˆ‚uâ€™/âˆ‚h)]. (âˆ‚h/âˆ‚u).( âˆ‚u/âˆ‚W_{i})â€¦â€¦â€¦â€¦â€¦(2)

So, What was the benefit of first calculating the gradient between the hidden layer and the output layer?

As you can see in equation (2) we have already computed âˆ‚E/âˆ‚Y and âˆ‚Y/âˆ‚uâ€™ saving us space and computation time. We will come to know in a while why is this algorithm called the back propagation algorithm.

Let us compute the unknown derivatives in equation (2).

âˆ‚uâ€™/âˆ‚h = âˆ‚(W_{h}h)/ âˆ‚h = W_{h}

âˆ‚h/âˆ‚u = âˆ‚( Ïƒ(u)/ âˆ‚u= Ïƒ(u)(1- Ïƒ(u))

But, Ïƒ(u)=h, So,

(âˆ‚Y/âˆ‚u)=h(1-h)

Now, âˆ‚u/âˆ‚W_{i }= âˆ‚(W_{i}X)/ âˆ‚W_{i} = X

Replacing all these values in equation (2) we get,

**âˆ‚E/âˆ‚W _{i} = [(Y-t). Y(1-Y).W_{h}].h(1-h).X**

So, now since we have calculated both the gradients, the weights can be updated as

**W _{h} = W_{h} + **

**Î· . âˆ‚E/âˆ‚W**

_{h}**W _{i} = W_{i} + **

**Î· . âˆ‚E/âˆ‚W**

_{i}Where **Î· **is the learning rate.

So coming back to the question: Why is this algorithm called Back Propagation Algorithm?

The reason is: If you notice the final form of **âˆ‚E/âˆ‚W _{h} **and

**âˆ‚E/âˆ‚W**, you will see the term (Y-t) i.e the output error, which is what we started with and then propagated this back to the input layer for weight updation.

_{i }So, where does this mathematics fit into the code?

hiddenlayer_activations=h

E= Y-t

Slope_output_layer = Y(1-Y)

lr = Î·

slope_hidden_layer = h(1-h)

wout = W_{h}

Now, you can easily relate the code to the mathematics.

**End Notes:**

This article is focused on the building a Neural Network from scratch and understanding its basic concepts. I hope now you understand the working of a neural networkÂ like how does forward and backward propagation work, optimization algorithms (Full Batch and Stochastic gradient descent), Â how to update weights and biases, visualization of each step in Excel and on top of that code in python and R.

Therefore, in myÂ upcoming article, Iâ€™ll explain the applications of using Neural Network in Python and solving real-life challenges related to:

- Computer Vision
- Speech
- Natural Language Processing

I enjoyed writing this article and would love to learn from your feedback. DidÂ you findÂ this article useful? I would appreciate yourÂ suggestions/feedback.Â Please feel free to ask your questions through comments below.

Thank you very much.

Amazing article..

Very well written and easy to understand the basic concepts..

Thank you for the hard work.

Thanks, for sharing this. Very nice article.

Nice article Sunil! Appreciate your continued research on the same.

One correction though…

Now…

hiddenlayer_neurons = 3 #number of hidden layers

Should be…

hiddenlayer_neurons = 3 #number of neurons at hidden layers

Thanks Srinivas! Have updated the comment.

i didn’t understand what is the need to calculate delta during back propagation.can you give any explanation to it.

Very interesting!

Very well written… I completely agree with you about learning by working on a problem

Thanks for great article! Probably, it should be “Update bias at both output and hidden layer” in the Step 11 of the Visualization of steps for Neural Network methodology

Thanks Andrei,

I’m updating only biases at step 11.

Regards,

Sunil

Wonderful explanation. This is an excellent article. I did not come across such a lucid explanation of NN so far.

Thanks Sasikanth!

Regards,

Sunil

Great article! There is a small typo: In the section where you describe the three ways of creating input output relationships you define “x2” twice – one of them should be “x3” instead ðŸ™‚

Keep up the great work!

Thanks Robert for highlighting the typo!

Explained in very lucid manner. Thanks for this wonderful article.

Very Interesting! Nice Explanation

Awesome Sunil. Its a great job.

Thanks a lot for making such a neat and clear page for NN, very much useful for beginners.

Well written article. With step by step explaination , it was easier to understand forward and backward propogations.. is there any functions in scikit learn for neural networks?

Thanks Praveen! You can look at this (http://scikit-learn.org/stable/modules/classes.html#module-sklearn.neural_network).

Regards,

Sunil

Hello Sunil,

Please refer below,

“To get a mathematical perspective of the Backward propagation, refer below section.

This one round of forward and back propagation iteration is known as one training iteration aka â€œEpochâ€œ. ”

I’m kind of lost there, did you already explain something?( about back prop) , Is there any missing information?

Thanks

Great article Sunil! I have one doubt. Why you applied linear to nonlinear transformation in the middle of the process? Is it necessary!!

Thanks a lot, Sunil, for such a well-written article. Particularly, I liked the visualization section, in which each step is well explained by an example.

I just have a suggestion: if you add the architecture of MLP in the beginning of the visualization section it would help a lot. Because in the beginning I thought you are addressing the same architecture plotted earlier, in which there were 2 hidden units, not 3 hidden units.

Thanks a lot once more!

Nice Article ðŸ™‚

Very well written article. Thanks for your efforts.

Great article. For a beginner like me, it was fully understandable. Keep up the good work.

Great Explanation….on Forward and Backward Propagation

Thanks Preeti

Regards,

Sunil

I really like how you explain this. Very well written. Thank you

Thanks Gino

I am 63 years old and retired professor of management. Thanks for your lucid explanations. I am able to learn. My blessings are to you.

Thanks Professor

Regards,

Sunil

Dear Author this is a great article. Infact I got more clarity.

I just wanted to say, using full batch Gradient Descent (or SGD) we need to tune the learning rate as well, but if we use Nesterovs Gradient Descent, it would converge faster and produce quick results.

good information thanks sunil

Hey sunil,

Can you also follow up with an article on rnn and lstm, with your same visual like tabular break down? It was fun and would complement a good nn understanding.

Thanks

A pleasant reading.

Thanks for sharing.

Thanks for the detailed explanation!

I want to hug you. I still have to read this again but machine learning algorithms have been shrouded in mystery before seeing this article. Thank you for unveiling it good friend.

Nice one.. Thanks lot for the work. i understood the neural network in a day

Yes, I found the information helpful in I understanding Neural Networks, I have and old book on the subject,

the book I found was very hard to understand, I enjoyed reading most of your article, I found how you presented the information good, I understood the language you used in writing the material, Good Job!

Thanks for great article, it is useful to understand the basic learning about neural networks. Thnaks again for making great effort…

benefit a lot

Thank you for this excellent plain-English explanation for amateurs.

Thank you, sir, very easy to understand and easy to practice.

Wonderful inspiration and great explanation.

Thank you very much

That is the simplest explain which i saw. Thx!

Thanks for the explanations, very clear

well done ðŸ˜€

A unique approach to visualize MLP ! Thank you …

I’m a beginner of this way. This article makes me understand about neural better. Thank you very much.

This is awesome explanation Sunil. The code and excel illustrations help a lot with really understanding the implementation. This helps unveil the mystery element from neural networks.

Thank you so much. This is what i wanted to know about NN.

Visualization is really very helpful. Thanks

Great article. The way of explanation is unbelievable. Thank you for writing.

Appreciate…. Stay Blessed.

Thanks this was a very good read.

Simply brilliant. Very nice piecemeal explanation. Thank you

very clear! thank you!

Thank you for your article. I have learned lots of DL from it.

Thank you very much. Very simple to understand ans easy to visualize. Please come up with more articles. Keep up the good work!

amazing article

thank you very much !!!!

This is amazing Mr. Sunil. Although am not a professional but a student, this article was very helpful in understanding the concept and an amazing guide to implement neural networks in python.

Mr. Sunil,

This was a great write-up and greatly improved my understanding of a simple neural network.

In trying to replicate your Excel implementation, however, I believe I found an error in Step 6, which calculates the output delta. What you have highlighted is the derivative of the Sigmoid function acting on the first column of the output_layer_input (not shown in image), and not on the actual output, which is what should actually happen and does happen in your R and Python implementations.

Thanks again!

Very well explanation. Everywhere NN is implemented using different libraries without defining fundamentals. Thanks a lot……

Very Simple Way But Best Explanation.

Thank You very much for explaining the concepts in a simple way.