Description
The network should implement the XOR function. Perform one epoch of backpropagation as introduced in the lecture on multilayer perceptrons.
Notes:
β’ The activation function f for a perceptron is the sigmoid function:
1
π(π₯)=1+π!”
β’ The thresholds are not shown in the network. The threshold nodes are set to -1.
β’ Use the following initial parameter values:
π#$($) =2 π#β($) =β1 π#$(β) =β2
π$$($) =6 π$β($) =8 π$$(β) =6
πβ$($) =β6 πββ($) =β8 πβ$(β) =β6
β’ The learning rate is set to Ξ· = 0.7
i. Compute the activations of the hidden and output neurons.
ii. Compute the error of the network.
iii. Backpropagate the error to determine βΞΈij for all weights ΞΈij and updates the weight
ΞΈij.
2. What is the difference between βmodel biasβ and βmodel varianceβ?
i. Why is a high bias, low variance classifier undesirable? ii. Why is a low bias, high variance classifier (usually) undesirable?
3. Describe how validation set, and cross-validation can help reduce overfitting?
Reviews
There are no reviews yet.