Size of Weight Matrices in an MLP Model

In a simple MLP model with 6 neurons in input layers, 4 neurons in the hidden layers, and 1 neuron in the output layer

What is the size of the weight matrices between the hidden output layer and input hidden layer?

a) [1X4], [4x6]

b) [6x4], [1x4]

c) [6x4], [4x1]

d) [4x1], [6x4]

The size of the weight matrices is [4x1] and [6x4]. hence the answer is option d).

To determine the size of the weight matrices between the hidden-output layer and the input-hidden layer in a simple MLP model, we need to consider the number of neurons in each layer. In this case, the hidden layer has 4 neurons and the input layer has 6 neurons. In conclusion, the weight matrix between the hidden-output layer will have a size of [4x1], and the weight matrix between the input-hidden layer will have a size of [6x4].

What is the size of the weight matrices between the hidden output layer and input hidden layer in a simple MLP model? The size of the weight matrices between the hidden output layer and input hidden layer is [4x1] and [6x4].
← Nosql database terminologies and features explained Creating reflective content with data →