Starting from:

$30

Cmput 466 Assignment 5

Problem 1.
Consider the training objective ๐ฝ = ||๐‘‹๐‘ค − ๐‘ก|| subject to for some constant .
2
||๐‘ค||
2 ≤ ๐ถ ๐ถ
How would the hypothesis class capacity, overfitting/underfittting, and bias/variance vary
according to ๐ถ?
Larger ๐ถ Smaller ๐ถ
Model capacity (large/small?) _____ _____
Overfitting/Underfitting? __fitting __fitting
Bias variance (how/low?) __ bias / __ variance __ bias / __ variance
Note: No proof is needed
Problem 2.
Consider a one-dimensional linear regression model ๐‘ก with a Gaussian prior
(๐‘š) ∼ ๐‘(๐‘ค๐‘ฅ
(๐‘š)
, σ
ฯต
2
)
๐‘ค ∼ ๐‘(0, σ . Show that the posterior of is also a Gaussian distribution, i.e., ๐‘ค
2
) ๐‘ค
๐‘ค|๐‘ฅ . Give the formulas for .
(1)
, ๐‘ก
(1)
, ···, ๐‘ฅ
(๐‘€)
, ๐‘ก
(๐‘€) ∼ ๐‘(µ
๐‘๐‘œ๐‘ ๐‘ก
, σ
๐‘๐‘œ๐‘ ๐‘ก
2
) µ
๐‘๐‘œ๐‘ ๐‘ก
, σ
๐‘๐‘œ๐‘ ๐‘ก
2
Hint: Work with ๐‘ƒ(๐‘ค|๐ท) ∝ ๐‘ƒ(๐‘ค)๐‘ƒ(๐ท|๐‘ค). Do not handle the normalizing term.
Note: If a prior has the same formula (but typically with different parameters) as the posterior, it
is known as a conjugate prior. The above conjugacy also applies to multi-dimensional Gaussian,
but the formulas for the mean vector and the covariance matrix will be more complicated.
Problem 3.
Give the prior distribution of ๐‘ค for linear regression, such that the max a posteriori estimation is
equivalent to ๐‘™ -penalized mean square loss.
1
Note: Such a prior is known as the Laplace distribution. Also, getting the normalization factor in
the distribution is not required.
END OF W5

More products