Q1
Q1
The McCulloch-Pitts neural model, which was the earliest ANN model, has only two types
of inputs — Excitatory and Inhibitory. The excitatory inputs have weights of positive
magnitude and the inhibitory weights have weights of negative magnitude. The inputs of
the McCulloch-Pitts neuron could be either 0 or 1. It has a threshold function as an
activation function. So, the output signal yout is 1 if the input ysum is greater than or equal to a
given threshold value, else 0. The diagrammatic representation of the model is as follows:
Simple McCulloch-Pitts neurons can be used to design logical operations. For that purpose,
the connection weights need to be correctly decided along with the threshold function
(rather than the threshold value of the activation function). For better understanding
purpose, let me consider an example:
John carries an umbrella if it is sunny or if it is raining. There are four given situations. I
need to decide when John will carry the umbrella. The situations are as follows:
First scenario: It is not raining, nor it is sunny
Second scenario: It is not raining, but it is sunny
Third scenario: It is raining, and it is not sunny
Fourth scenario: It is raining as well as it is sunny
To analyse the situations using the McCulloch-Pitts neural model, I can consider the input
signals as follows:
X1: Is it raining?
X2 : Is it sunny?
So, the value of both scenarios can be either 0 or 1. We can use the value of both weights
X1 and X2 as 1 and a threshold function as 1. So, the neural network model will look like:
So, I can say that, The truth table built with respect to the problem is depicted above. From
the truth table, I can conclude that in the situations where the value of yout is 1, John needs
to carry an umbrella. Hence, he will need to carry an umbrella in scenarios 2, 3 and 4.
The delta rule is an update rule for single layer perceptrons. It makes use of gradient descent.
Delta rule in neural networks that is used to updated weights during training a Neural
Network.
Introduction
The delta rule is a formula for updating the weights of a neural network during training. It is
considered a special case of the backpropagation algorithm. The delta rule is in fact a gradient
descent learning rule.
Recall that the process of training a neural network involves iterating over the following
steps:
A set of input and output sample pairs are selected randomly and run through the neural
network. The network makes predictions on these samples.
The loss between the predictions and the true values is computed.
Application
The generalized delta rule is important in creating useful networks capable of learning
complex relations between inputs and outputs. Compared to other early learning rules like the
Hebbian learning rule or the Correlation learning rule, the delta rule has the advantage that it
is mathematically derived and directly applicable to Supervised Learning. In addition, unlike
the Perceptron learning rule which relies on the use of the Heaviside step function as the
activation function which means that the derivative does not exist at zero, and is equal to zero
elsewhere, the delta rule is applicable to differentiable activation functions like tanh and the
sigmoid function.
Like a single-layer perceptron model, a multi-layer perceptron model also has the same
model structure but has a greater number of hidden layers.
The multi-layer perceptron model is also known as the Backpropagation algorithm, which
executes in two stages as follows:
o Forward Stage: Activation functions start from the input layer in the forward stage
and terminate on the output layer.
o Backward Stage: In the backward stage, weight and bias values are modified as per
the model's requirement. In this stage, the error between actual output and demanded
originated backward on the output layer and ended on the input layer.
Hence, a multi-layered perceptron model has considered as multiple artificial neural networks
having various layers in which activation function does not remain linear, similar to a single
layer perceptron model. Instead of linear, activation function can be executed as sigmoid,
TanH, ReLU, etc., for deployment.
A multi-layer perceptron model has greater processing power and can process linear and non-
linear patterns. Further, it can also implement logic gates such as AND, OR, XOR, NAND,
NOT, XNOR, NOR.
5. Scalability: Backpropagation scales well with the size of the dataset and
the complexity of the network. This scalability makes it suitable for large-
scale machine learning tasks, where training data and network size are
significant factors.
In conclusion, the backpropagation algorithm offers several advantages that contribute to its
widespread use in training neural networks. Its ease of implementation, simplicity, efficiency,
generalization ability, and scalability make it a valuable tool for developing and training
neural network models for various machine learning applications.