Suppose you have built a neural network. You decide to initialize the weights and biases to be zero. Which of the following statements are True? (Check all that apply)
Out of the given options, what is the best choice for number of clusters(k) ?
In general if you were given a choice of activation function where both Sigmoid and Leaky ReLu can be used, which one would you tend to prefer?
The network that involves backward links from output to the input and hidden layers is called?