Unit6 - Subjective Questions

INT255 • Practice Questions with Detailed Answers

1

Define bias and variance in the context of machine learning models. How do these two components contribute to the overall generalization error?

2

Explain the concept of the bias-variance trade-off. Illustrate this trade-off with a conceptual graph showing training error, test error, bias, and variance as model complexity changes.

3

Provide the mathematical decomposition of the expected generalization error into bias, variance, and irreducible error components. Briefly explain each component.

4

Distinguish between overfitting and underfitting from a practical and mathematical viewpoint. Describe their typical signs during model training and validation.

5

Using polynomial regression as an example, explain how varying the degree of the polynomial can lead to underfitting or overfitting.

6

Explain L1 regularization (Lasso) in the context of linear models. What is its primary effect on model coefficients, and why is it useful for feature selection?

7

Explain L2 regularization (Ridge) in the context of linear models. How does it differ from L1 regularization in its effect on model coefficients?

8

Write down the objective functions for:

  1. Ordinary Least Squares (OLS) regression.
  2. Ridge regression (L2 regularization).
  3. Lasso regression (L1 regularization).

Explain the purpose of the regularization terms in the latter two.

9

Compare and contrast L1 and L2 regularization. Discuss their respective advantages, disadvantages, and typical use cases.

10

Describe the geometric interpretation of L1 and L2 regularization. Explain how the shapes of their constraint regions (L1-ball and L2-ball) lead to different properties, especially sparsity for L1.

11

How does the regularization parameter (lambda) control the balance between fitting the training data and keeping the model simple? Relate this to the bias-variance trade-off.

12

Explain the concept of "norm-based constraints" in regularization. How do L1 and L2 regularization fit this description, and what is their role in preventing overfitting?

13

Explain the fundamental principle of Structural Risk Minimization (SRM). How does it attempt to improve upon Empirical Risk Minimization (ERM) for better generalization?

14

According to the SRM framework, what are the two main components that determine the generalization error bound of a model? Briefly explain how SRM aims to minimize this bound.

15

Discuss how the concept of Structural Risk Minimization (SRM) provides a theoretical justification for using regularization techniques in machine learning.

16

What is the Vapnik-Chervonenkis (VC) dimension? Provide an intuitive explanation of what it measures regarding a hypothesis class.

17

Explain the concept of "shattering" in the context of VC dimension. Provide an intuitive example to illustrate how a hypothesis class can shatter a set of points (e.g., using linear classifiers on points in 1D or 2D).

18

How does a higher VC dimension relate to the complexity of a model and its capacity to fit arbitrary data patterns?

19

Explain the role of VC dimension in generalization theory. How does it influence the bounds on generalization error?

20

Beyond regularization, list and briefly describe at least three other strategies or techniques used to mitigate overfitting and underfitting in machine learning models.