1What is the primary goal of Simple Linear Regression?
A.To classify data into discrete categories
B.To find the relationship between two continuous variables using a straight line
C.To group similar data points together
D.To reduce the dimensionality of the dataset
Correct Answer: To find the relationship between two continuous variables using a straight line
Explanation:Simple Linear Regression aims to model the relationship between a single independent variable and a dependent variable by fitting a linear equation to observed data.
Incorrect! Try again.
2In the linear regression equation Y = mx + c, what does 'c' represent?
A.The slope of the line
B.The dependent variable
C.The y-intercept
D.The error term
Correct Answer: The y-intercept
Explanation:In the equation Y = mx + c, 'c' is the y-intercept, which represents the value of Y when x is 0.
Incorrect! Try again.
3Which method is commonly used to estimate the coefficients in linear regression?
A.Maximum Likelihood Estimation
B.Ordinary Least Squares (OLS)
C.K-Means Clustering
D.Principal Component Analysis
Correct Answer: Ordinary Least Squares (OLS)
Explanation:Ordinary Least Squares (OLS) is the most common method used to estimate the parameters of a linear regression model by minimizing the sum of the squares of the differences between observed and predicted values.
Incorrect! Try again.
4What is a 'residual' in the context of regression?
A.The value of the independent variable
B.The difference between the observed value and the predicted value
C.The slope of the regression line
D.The square of the correlation coefficient
Correct Answer: The difference between the observed value and the predicted value
Explanation:A residual is the vertical distance between a data point and the regression line, calculated as (Actual Value - Predicted Value).
Incorrect! Try again.
5Which of the following values indicates a perfect negative correlation?
A.
B.1
C.-1
D.0.5
Correct Answer: -1
Explanation:A Pearson correlation coefficient of -1 indicates a perfect negative linear relationship between two variables.
Incorrect! Try again.
6Multiple Linear Regression differs from Simple Linear Regression because it involves:
A.Multiple dependent variables
B.Multiple independent variables
C.Polynomial terms
D.Categorical outputs
Correct Answer: Multiple independent variables
Explanation:Multiple Linear Regression models the relationship between two or more independent variables (features) and a single dependent variable.
Incorrect! Try again.
7In a regression model, if the R-squared value is 0.85, what does this mean?
A.85% of the errors are zero
B.The model is 85% accurate in classification
C.85% of the variance in the dependent variable is explained by the independent variables
D.The correlation coefficient is 0.85
Correct Answer: 85% of the variance in the dependent variable is explained by the independent variables
Explanation:R-squared represents the proportion of the variance for a dependent variable that's explained by an independent variable or variables in a regression model.
Incorrect! Try again.
8Which evaluation metric is calculated by taking the average of the squared differences between predicted and actual values?
A.MAE
B.RMSE
C.MSE
D.R-squared
Correct Answer: MSE
Explanation:Mean Squared Error (MSE) measures the average of the squares of the errors—that is, the average squared difference between the estimated values and the actual value.
Incorrect! Try again.
9Which regression algorithm is specifically designed to predict the probability of a categorical outcome (e.g., Yes/No)?
A.Simple Linear Regression
B.Polynomial Regression
C.Logistic Regression
D.Ridge Regression
Correct Answer: Logistic Regression
Explanation:Logistic Regression is used for binary classification problems to predict the probability of an instance belonging to a specific class.
Incorrect! Try again.
10What function is used in Logistic Regression to map predicted values to probabilities between 0 and 1?
A.Linear function
B.Sigmoid (Logistic) function
C.ReLU function
D.Polynomial function
Correct Answer: Sigmoid (Logistic) function
Explanation:The Sigmoid function maps any real-valued number into a value between 0 and 1, making it suitable for probability estimation in Logistic Regression.
Incorrect! Try again.
11Polynomial Regression is best used when:
A.The relationship between variables is linear
B.The relationship between variables is non-linear
C.The dependent variable is categorical
D.There are too many independent variables
Correct Answer: The relationship between variables is non-linear
Explanation:Polynomial Regression models a non-linear relationship between the independent and dependent variables by adding powers of the independent variable.
Incorrect! Try again.
12What is a potential risk of using a high-degree polynomial in Polynomial Regression?
A.Underfitting
B.Overfitting
C.Linearity
D.High bias
Correct Answer: Overfitting
Explanation:Using a high-degree polynomial allows the model to fit the training data very closely, often capturing noise and fluctuations, leading to overfitting and poor generalization to new data.
Incorrect! Try again.
13Root Mean Squared Error (RMSE) is preferred over MSE when:
A.You want the error metric to be in the same units as the target variable
B.You want to penalize outliers less
C.You have a classification problem
D.You want a value between 0 and 1
Correct Answer: You want the error metric to be in the same units as the target variable
Explanation:RMSE is the square root of MSE, which brings the error metric back to the original units of the target variable, making it more interpretable.
Incorrect! Try again.
14Which metric is less sensitive to outliers?
A.MSE
B.RMSE
C.MAE
D.R-squared
Correct Answer: MAE
Explanation:Mean Absolute Error (MAE) takes the absolute difference without squaring, meaning large errors (outliers) do not contribute disproportionately to the total error compared to MSE or RMSE.
Incorrect! Try again.
15In OLS estimation, the assumption of 'Homoscedasticity' implies that:
A.The residuals have constant variance across all levels of the independent variable
B.The residuals follow a normal distribution
C.There is no correlation between independent variables
D.The relationship is non-linear
Correct Answer: The residuals have constant variance across all levels of the independent variable
Explanation:Homoscedasticity means that the spread (variance) of the residuals is consistent for all predicted values. If the spread changes (e.g., cone shape), it is Heteroscedasticity.
Incorrect! Try again.
16Multicollinearity in Multiple Linear Regression refers to:
A.A high correlation between the dependent and independent variables
B.A high correlation between two or more independent variables
C.The lack of a linear relationship
D.Measurement errors in the target variable
Correct Answer: A high correlation between two or more independent variables
Explanation:Multicollinearity occurs when independent variables in a regression model are highly correlated with each other, making it difficult to isolate the individual effect of each variable.
Incorrect! Try again.
17The range of the R-squared (R²) score is typically:
A.-1 to 1
B.0 to infinity
C.0 to 1
D.-infinity to 1
Correct Answer: 0 to 1
Explanation:R-squared typically ranges from 0 to 1, where 1 indicates a perfect fit. However, it can be negative for arbitrarily poor models that perform worse than a horizontal mean line.
Incorrect! Try again.
18In the equation Y = b0 + b1x1 + b2x2 + ... + bn*xn, what are b1, b2, ... bn called?
A.Intercepts
B.Residuals
C.Regression coefficients
D.Dependent variables
Correct Answer: Regression coefficients
Explanation:These are the regression coefficients (or weights) that represent the change in the dependent variable for a one-unit change in the respective independent variable.
Incorrect! Try again.
19Which of the following is NOT an assumption of Linear Regression?
A.Linearity
B.Independence of errors
C.Multicollinearity usually present
D.Normality of residuals
Correct Answer: Multicollinearity usually present
Explanation:Linear Regression assumes little to no multicollinearity. The presence of significant multicollinearity is a violation of assumptions.
Incorrect! Try again.
20What is the primary advantage of Adjusted R-squared over R-squared?
A.It is easier to calculate
B.It penalizes the addition of irrelevant independent variables
C.It can handle non-linear data
D.It is always higher than R-squared
Correct Answer: It penalizes the addition of irrelevant independent variables
Explanation:Regular R-squared never decreases when new variables are added, even if they are useless. Adjusted R-squared accounts for the number of predictors and decreases if a new term improves the model less than expected by chance.
Incorrect! Try again.
21In Logistic Regression, the 'Logit' is defined as:
A.The sum of squared errors
B.The natural logarithm of the odds ratio
C.The square root of the variance
D.The probability of success
Correct Answer: The natural logarithm of the odds ratio
Explanation:The Logit function is the log-odds, defined as log(p / (1-p)), which transforms the probability to a value ranging from negative to positive infinity, suitable for linear modeling.
Incorrect! Try again.
22If the Pearson correlation coefficient between X and Y is 0, it implies:
A.X and Y are identical
B.There is no linear relationship between X and Y
C.X causes Y
D.There is a strong non-linear relationship
Correct Answer: There is no linear relationship between X and Y
Explanation:A correlation of 0 indicates no linear relationship, though a non-linear relationship could still exist.
Incorrect! Try again.
23Which cost function is primarily used for Logistic Regression?
A.Mean Squared Error
B.Mean Absolute Error
C.Log Loss (Cross-Entropy)
D.Hinge Loss
Correct Answer: Log Loss (Cross-Entropy)
Explanation:Log Loss (or Binary Cross-Entropy) is the standard cost function for Logistic Regression because using MSE would result in a non-convex function that is hard to optimize.
Incorrect! Try again.
24When evaluating a regression model, a lower MAE indicates:
A.Worse performance
B.Better performance
C.High variance
D.Overfitting
Correct Answer: Better performance
Explanation:MAE measures the average magnitude of errors. A value closer to 0 indicates the predictions are closer to the actual values.
Incorrect! Try again.
25To handle categorical independent variables in regression, one should usually:
A.Ignore them
B.Convert them using One-Hot Encoding (Dummy variables)
C.Assign random numbers
D.Use the text directly
Correct Answer: Convert them using One-Hot Encoding (Dummy variables)
Explanation:Regression models require numerical input. Categorical variables are converted into dummy variables (0s and 1s) to be processed.
Incorrect! Try again.
26The 'Dummy Variable Trap' occurs when:
A.Categorical variables are missing
B.One dummy variable can be predicted perfectly from the others (perfect multicollinearity)
C.There are too many categories
D.The variables are not scaled
Correct Answer: One dummy variable can be predicted perfectly from the others (perfect multicollinearity)
Explanation:This happens if you include a dummy variable for every category. The sum of all dummy variables equals 1 (the intercept term), causing perfect multicollinearity. One category should be dropped to avoid this.
Incorrect! Try again.
27Which of the following plots is best for visualizing a Simple Linear Regression?
A.Pie Chart
B.Scatter Plot with a line of best fit
C.Histogram
D.Box Plot
Correct Answer: Scatter Plot with a line of best fit
Explanation:A scatter plot displays the relationship between two continuous variables, and the line of best fit visualizes the regression model.
Incorrect! Try again.
28If a polynomial regression model has degree 1, it behaves like:
A.A Simple Linear Regression
B.A Logistic Regression
C.A Quadratic Regression
D.An Exponential Regression
Correct Answer: A Simple Linear Regression
Explanation:A polynomial of degree 1 is simply y = b0 + b1*x, which is the equation for a straight line (Simple Linear Regression).
Incorrect! Try again.
29In the context of MSE, what is the effect of squaring the errors?
A.It cancels out positive and negative errors
B.It penalizes larger errors more severely than smaller errors
C.It makes the calculation faster
D.It reduces the impact of outliers
Correct Answer: It penalizes larger errors more severely than smaller errors
Explanation:Squaring the difference ensures that all errors are positive and that larger errors contribute disproportionately more to the total score, forcing the model to minimize large deviations.
Incorrect! Try again.
30Which algorithm minimizes the sum of squared residuals?
A.Gradient Boosting
B.Ordinary Least Squares
C.Logistic Regression
D.Decision Trees
Correct Answer: Ordinary Least Squares
Explanation:The fundamental mathematical principle behind standard linear regression is OLS, which explicitly minimizes the sum of squared residuals.
Incorrect! Try again.
31What is the typical threshold used in Logistic Regression to classify a probability as '1' (or Positive)?
A.0.1
B.1.0
C.0.5
D.0.0
Correct Answer: 0.5
Explanation:While adjustable, the standard default threshold is 0.5. Probabilities >= 0.5 are classified as positive (1), and < 0.5 as negative (0).
Incorrect! Try again.
32Predicting the price of a house based on size, location, and age is a problem of:
A.Clustering
B.Classification
C.Regression
D.Dimensionality Reduction
Correct Answer: Regression
Explanation:Predicting a continuous numerical value (price) is a Regression task.
Incorrect! Try again.
33Predicting whether an email is 'Spam' or 'Not Spam' is a problem of:
Explanation:Predicting a categorical outcome (binary in this case) is a classification task, for which Logistic Regression is suitable.
Incorrect! Try again.
34In the equation Y = b0 + b1*X + e, what does 'e' represent?
A.The predicted value
B.The error term (noise)
C.The correlation
D.The intercept
Correct Answer: The error term (noise)
Explanation:The term 'e' (epsilon) represents the error term, accounting for the variance in Y that is not explained by the linear relationship with X.
Incorrect! Try again.
35If the correlation coefficient (r) is 0.9, the Coefficient of Determination (R²) is:
A.0.9
B.0.18
C.0.81
D.0.45
Correct Answer: 0.81
Explanation:In simple linear regression, R-squared is the square of the Pearson correlation coefficient ().
Incorrect! Try again.
36Which of the following indicates the strongest relationship?
A.Correlation = 0.5
B.Correlation = -0.8
C.Correlation = 0.1
D.Correlation = 0.7
Correct Answer: Correlation = -0.8
Explanation:The strength of a relationship is determined by the absolute value of the correlation. |-0.8| = 0.8, which is higher than 0.7 or 0.5.
Incorrect! Try again.
37Why might one perform a log transformation on the dependent variable in regression?
A.To make the distribution more normal or linearize a relationship
B.To turn it into a categorical variable
C.To increase the number of outliers
D.To reduce the sample size
Correct Answer: To make the distribution more normal or linearize a relationship
Explanation:Log transformations are often used to handle skewed data or to transform a non-linear exponential relationship into a linear one.
Incorrect! Try again.
38If a model has an RMSE of 10 and an MAE of 2, what does this suggest?
A.The model is perfect
B.There are likely large outliers in the errors
C.The model is underfitting
D.RMSE is calculated incorrectly
Correct Answer: There are likely large outliers in the errors
Explanation:Because RMSE squares errors, it is much more sensitive to outliers than MAE. A large gap between MAE and RMSE suggests the presence of large error values (outliers).
Incorrect! Try again.
39In polynomial regression, if the curve passes through every single data point perfectly, the model is likely:
A.Ideally fitted
B.Underfitted
C.Overfitted
D.Linear
Correct Answer: Overfitted
Explanation:Fitting every point perfectly usually means the model has captured the noise rather than the underlying pattern, leading to overfitting.
Incorrect! Try again.
40The decision boundary in Logistic Regression is:
A.Curved
B.Linear (in the feature space)
C.Circular
D.Undefined
Correct Answer: Linear (in the feature space)
Explanation:Standard Logistic Regression creates a linear decision boundary that separates the classes in the feature space.
Incorrect! Try again.
41What is the slope of the line y = 3x + 5?
A.5
B.3
C.
D.8
Correct Answer: 3
Explanation:In the slope-intercept form , is the slope. Here, .
Incorrect! Try again.
42When comparing two regression models on the same dataset, the one with the ___ is generally preferred.
A.Higher RMSE
B.Lower R-squared
C.Higher R-squared and Lower RMSE
D.Higher MAE
Correct Answer: Higher R-squared and Lower RMSE
Explanation:A better model explains more variance (higher R²) and has smaller average errors (lower RMSE).
Incorrect! Try again.
43Which technique helps checks for 'Linearity' in a regression model?
A.Residuals vs Predicted Values plot
B.Histogram of the target
C.Correlation matrix
D.Bar chart
Correct Answer: Residuals vs Predicted Values plot
Explanation:A plot of residuals versus predicted values should show a random scatter. If there is a pattern (like a curve), the linearity assumption is violated.
Incorrect! Try again.
44The 'Bias-Variance Tradeoff' implies that:
A.We want high bias and high variance
B.As we increase model complexity, bias decreases but variance increases
C.Simple models have high variance
D.Complex models have high bias
Correct Answer: As we increase model complexity, bias decreases but variance increases
Explanation:Simple models (high bias) miss patterns. Complex models (high variance) overfit. The goal is to find the sweet spot.
Incorrect! Try again.
45Feature scaling (normalization/standardization) is particularly important for:
A.Simple Linear Regression with one variable
B.Decision Trees
C.Regression using Gradient Descent optimization
D.Ordinary Least Squares solution
Correct Answer: Regression using Gradient Descent optimization
Explanation:While OLS is scale-invariant analytically, optimization algorithms like Gradient Descent converge much faster when features are on a similar scale.
Incorrect! Try again.
46If R-squared is 1.0:
A.The model predicts the mean for every observation
B.The model explains 100% of the variability in the data
C.The model is completely wrong
D.The correlation is 0
Correct Answer: The model explains 100% of the variability in the data
Explanation:An R-squared of 1.0 means the model perfectly fits the data, and the regression line passes through all observed points.
Incorrect! Try again.
47Which of the following is a dependent variable in a study of how study time affects exam scores?
A.Study time
B.Exam scores
C.Student ID
D.Subject
Correct Answer: Exam scores
Explanation:The dependent variable is the outcome we are trying to predict. Here, exam scores depend on the study time.
Incorrect! Try again.
48In the context of regression, 'extrapolation' refers to:
A.Predicting values outside the range of the training data
B.Predicting values within the range of training data
C.Removing outliers
D.Calculating the mean
Correct Answer: Predicting values outside the range of the training data
Explanation:Extrapolation involves making predictions for independent variable values that are outside the range of data used to train the model, which is risky.
Incorrect! Try again.
49The correlation coefficient ranges between:
A.0 and 1
B.-1 and 1
C.-infinity and infinity
D.0 and 100
Correct Answer: -1 and 1
Explanation:Pearson's correlation coefficient is always between -1 (perfect negative) and +1 (perfect positive).
Incorrect! Try again.
50Which of the following is NOT a metric for Regression?
A.Accuracy
B.MAE
C.MSE
D.R-squared
Correct Answer: Accuracy
Explanation:Accuracy is a classification metric (percentage of correct classes). It does not apply to regression, where errors are continuous values.
Incorrect! Try again.
Give Feedback
Help us improve by sharing your thoughts or reporting issues.