This article was originally published at Codeperfectplus.
The study of the human brain is thousands of years old. The first step toward artificial neural networks came in 1943 when Warren McCulloch, a neurophysiologist, and a young mathematician, Walter Pitts, wrote a paper on how neurons might work. They modelled a simple neural network with electrical circuits.
Perceptron is a binary linear classification algorithm. It's a supervised type of machine learning and the simplest form of neural network. It's a base for neural networks. So, if you want to know how neural network works, learn how perception works.
These are the major components of perceptrons.
- Input Value
- Weight and Bias
- Net Sum
- Activation Function
Input Value:- Input Values are the feature of the Data set.
Weight and Bias:- Weights are the values that are computed over the time of training the model. Initial we start the value of weights with some initial value and these values get updated for each training basis of the error function.
Bias is like the intercept added in a linear equation. It is an additional parameter in the Neural Network which is used to adjust the output along with the weighted sum of the inputs to the neuron.It helps in training the model faster and with better quality.
Net Sum:- The net sum is the sum of the multiplication of input and weight.
Activation Function:- Activation function is used to make neural network non-linear. It's limit the output between 0 and 1.
Any comments or if you have any question, write it in the comment.
Happy to be helpful.
- Deploy Your First Django App With Heroku
- Single-layer Neural Networks in Machine Learning (Perceptrons)
- 5 Tips for Computer Programming Beginners
- What Is Git and GitHub?
- What is Simple Linear Regression?
- Introduction to Machine Learning and it's Type.
- Difference Between Machine Learning and Artificial Intelligence