Unit3 - Subjective Questions

INT255 • Practice Questions with Detailed Answers

1

Define random variables and their significance in Machine Learning models. Differentiate between discrete and continuous random variables with examples relevant to ML.

2

Explain the Bernoulli and Binomial distributions. Provide examples of where each might be applied in a machine learning context.

3

Describe the Gaussian (Normal) distribution. Why is it so prevalent in machine learning, particularly in models like Linear Regression and Gaussian Mixture Models?

4

What is likelihood in the context of machine learning? Explain the principle of Maximum Likelihood Estimation (MLE) and its goal.

5

Derive the Maximum Likelihood Estimator for the parameter of a Bernoulli distribution given a dataset where each .

6

Define the squared error loss function. For what type of machine learning problems is it typically used, and why? Discuss its properties, including convexity.

7

Explain the logistic loss function (also known as binary cross-entropy loss). For what type of machine learning problems is it primarily used? Provide its mathematical formula and explain its relation to probability.

8

State Bayes' Theorem. Explain the roles of the prior, likelihood, and posterior distributions in the Bayesian interpretation of learning models.

9

Compare and contrast Maximum Likelihood Estimation (MLE) and Maximum A Posteriori (MAP) estimation. Under what conditions might MAP be preferred over MLE?

10

Discuss how the concept of a random variable is fundamental to understanding the output of a classification model or the error in a regression model. Give specific examples.

11

Describe the Categorical distribution and its application in multi-class classification problems. How does it relate to the softmax function?

12

Explain why the log-likelihood is often maximized instead of the likelihood itself in machine learning algorithms. Discuss the mathematical advantages of working with log-likelihood.

13

What are the desirable properties of a good loss function? Discuss how squared error and logistic loss demonstrate some of these properties.

14

Outline the steps to derive the Maximum Likelihood Estimators for the mean () and variance () of a Gaussian distribution given a dataset . (You don't need to perform the full derivation, but explain the process).

15

Clearly distinguish between probability and likelihood. Use an example to illustrate when you would use each term in a machine learning context.

16

Describe the general process of Bayesian inference in machine learning. How does it update beliefs about model parameters as new data arrives?

17

Explain how Maximum A Posteriori (MAP) estimation can be viewed as a form of regularization in machine learning. Provide an example linking a common regularization technique to a specific prior distribution.

18

Explain the relationship between the logistic loss (binary cross-entropy) and Maximum Likelihood Estimation for a Bernoulli distribution. Show how minimizing logistic loss is equivalent to maximizing the log-likelihood of a Bernoulli model.

19

What is the concept of "expected loss" in decision theory? How does it relate to the selection of a model's parameters in a probabilistic framework?

20

Briefly describe the Dirichlet distribution and its role as a prior in Bayesian contexts, particularly for parameters of categorical or multinomial distributions.