{ "title": "Lasso Regression Mastery: 100 MCQs", "description": "A comprehensive set of 100 multiple-choice questions designed to teach and test your understanding of Lasso Regression, from basic linear regression concepts to advanced feature selection, sparsity, regularization, and real-world scenarios.", "questions": [ { "id": 1, "questionText": "What is the main goal of Lasso Regression?", "options": [ "To predict categories", "To find a straight-line relationship and perform feature selection", "To compress the data", "To find clusters in data" ], "correctAnswerIndex": 1, "explanation": "Lasso Regression adds an L1 penalty, which can shrink some coefficients to exactly zero, performing feature selection while modeling relationships." }, { "id": 2, "questionText": "Lasso Regression differs from Ridge Regression because it:", "options": [ "Uses L1 penalty instead of L2", "Does not require standardization", "Cannot handle multicollinearity", "Always increases training error" ], "correctAnswerIndex": 0, "explanation": "Lasso uses L1 regularization, which can shrink some coefficients to zero, unlike Ridge which uses L2 and shrinks all coefficients." }, { "id": 3, "questionText": "Scenario: You have a dataset with 20 features, some irrelevant. Which regression is suitable for automatic feature selection?", "options": [ "Ridge Regression", "Linear Regression", "Lasso Regression", "Polynomial Regression" ], "correctAnswerIndex": 2, "explanation": "Lasso’s L1 penalty can zero out irrelevant features, performing automatic feature selection." }, { "id": 4, "questionText": "In Lasso Regression, increasing the alpha parameter will:", "options": [ "Ignore correlated features", "Shrink more coefficients to zero", "Decrease training error and increase test error", "Remove the intercept automatically" ], "correctAnswerIndex": 1, "explanation": "Higher alpha strengthens the L1 penalty, increasing sparsity and setting more coefficients exactly to zero." }, { "id": 5, "questionText": "Why is feature standardization important in Lasso Regression?", "options": [ "It removes noise automatically", "Alpha is irrelevant", "Intercept is ignored otherwise", "L1 penalty depends on coefficient magnitude, which depends on feature scale" ], "correctAnswerIndex": 3, "explanation": "Without standardization, features with large scales dominate the penalty, leading to unfair coefficient shrinkage." }, { "id": 6, "questionText": "Scenario: Two features are highly correlated. Lasso Regression may:", "options": [ "Select one feature and zero the other", "Ignore both", "Shrink both equally like Ridge", "Fail to converge" ], "correctAnswerIndex": 0, "explanation": "L1 penalty tends to pick one correlated feature and set the other to zero, performing feature selection." }, { "id": 7, "questionText": "Scenario: Lasso Regression applied with alpha=0. Result?", "options": [ "Removes multicollinearity", "Performs feature selection", "Reduces variance", "Behaves like Linear Regression" ], "correctAnswerIndex": 3, "explanation": "Alpha=0 removes the L1 penalty, reducing Lasso to standard Linear Regression." }, { "id": 8, "questionText": "Scenario: Lasso Regression applied to polynomial features with high degree. Main effect?", "options": [ "Eliminates low-degree terms automatically", "Increases training error only", "Removes intercept", "Controls overfitting by zeroing some coefficients" ], "correctAnswerIndex": 3, "explanation": "Lasso shrinks some high-degree term coefficients to zero, reducing overfitting." }, { "id": 9, "questionText": "Scenario: Lasso Regression applied on dataset with missing values. Action?", "options": [ "Remove alpha", "Impute missing values before Lasso", "L1 penalty handles missing automatically", "Ignore missing values" ], "correctAnswerIndex": 1, "explanation": "Lasso requires complete data; missing values should be imputed or removed first." }, { "id": 10, "questionText": "Scenario: Lasso Regression applied on standardized dataset. Alpha too high. Result?", "options": [ "Intercept removed automatically", "Coefficients remain unchanged", "Low training error, high variance", "Many coefficients set to zero, potential underfitting" ], "correctAnswerIndex": 3, "explanation": "Excessive regularization shrinks many coefficients to zero, increasing bias." }, { "id": 11, "questionText": "Scenario: Lasso vs Ridge on dataset with irrelevant features. Observation?", "options": [ "Ridge eliminates features; Lasso does not", "Ridge shrinks coefficients but keeps them non-zero; Lasso may zero irrelevant features", "Both produce identical results", "Alpha irrelevant" ], "correctAnswerIndex": 1, "explanation": "Lasso performs feature selection while Ridge shrinks coefficients but retains all features." }, { "id": 12, "questionText": "Scenario: Lasso Regression applied to time-series data with lag features. Standardization importance?", "options": [ "Ensures fair penalty for all lag features", "Alpha irrelevant", "Intercept ignored otherwise", "Training error minimized automatically" ], "correctAnswerIndex": 0, "explanation": "L1 penalty affects coefficients fairly only if features are standardized." }, { "id": 13, "questionText": "Scenario: Lasso Regression applied to one-hot encoded categorical variables. Concern?", "options": [ "May zero some dummy variables while keeping others", "Intercept removed automatically", "No need for scaling", "Alpha irrelevant" ], "correctAnswerIndex": 0, "explanation": "Lasso may select one dummy feature and set others to zero, performing feature selection." }, { "id": 14, "questionText": "Scenario: Lasso Regression applied to high-dimensional dataset (features > samples). Advantage?", "options": [ "Training error minimal", "Can produce sparse model, eliminating irrelevant features", "Cannot converge", "Removes intercept" ], "correctAnswerIndex": 1, "explanation": "L1 penalty creates sparsity, which is helpful in high-dimensional problems." }, { "id": 15, "questionText": "Scenario: Lasso Regression applied with cross-validation. Purpose of cross-validation?", "options": [ "Eliminate correlated features", "Always minimize training error", "Select optimal alpha to balance bias-variance tradeoff", "Ignore feature scaling" ], "correctAnswerIndex": 2, "explanation": "Cross-validation identifies alpha that minimizes validation/test error for better generalization." }, { "id": 16, "questionText": "Scenario: Lasso Regression applied to noisy dataset. Observed smaller coefficients than Linear Regression. Why?", "options": [ "Alpha zero", "Noise ignored automatically", "L1 penalty shrinks some coefficients to zero, reducing variance", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Regularization reduces sensitivity to noise, shrinking coefficients and improving generalization." }, { "id": 17, "questionText": "Scenario: Lasso Regression vs Elastic Net. Main difference?", "options": [ "Elastic Net ignores feature selection", "Lasso uses L2", "Elastic Net uses L1 only", "Elastic Net combines L1 and L2, useful when correlated features exist" ], "correctAnswerIndex": 3, "explanation": "Elastic Net combines L1 and L2 penalties, balancing feature selection and coefficient shrinkage." }, { "id": 18, "questionText": "Scenario: Lasso Regression applied to correlated features. Observation?", "options": [ "Shrinks coefficients equally", "Tends to select one and zero others", "Fails to converge", "Removes intercept" ], "correctAnswerIndex": 1, "explanation": "L1 penalty promotes sparsity, often selecting one feature from correlated groups." }, { "id": 19, "questionText": "Scenario: Lasso Regression applied with very small alpha. Effect?", "options": [ "High sparsity", "Fails to converge", "Minimal regularization, behaves like Linear Regression", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Small alpha means weak L1 penalty; model behaves like Linear Regression with minimal feature selection." }, { "id": 20, "questionText": "Scenario: Lasso Regression applied to polynomial features. High-degree coefficients are large. Solution?", "options": [ "Increase alpha to shrink some coefficients to zero", "Ignore intercept", "Decrease alpha", "Remove low-degree terms" ], "correctAnswerIndex": 0, "explanation": "Higher alpha reduces overfitting by zeroing some large coefficients from high-degree terms." }, { "id": 21, "questionText": "Scenario: Lasso Regression applied to dataset with irrelevant features. Test error high. Solution?", "options": [ "Ignore irrelevant features", "Increase alpha excessively", "Alpha tuning via cross-validation, or consider Elastic Net", "Decrease alpha to zero" ], "correctAnswerIndex": 2, "explanation": "Cross-validation or Elastic Net can select relevant features and improve generalization." }, { "id": 22, "questionText": "Scenario: Lasso Regression applied to dataset with standardized features. Correlated features coefficients?", "options": [ "One selected, others may be zero", "Fails to converge", "All coefficients equal", "Intercept removed automatically" ], "correctAnswerIndex": 0, "explanation": "L1 penalty often selects one correlated feature and zeros others." }, { "id": 23, "questionText": "Scenario: Lasso Regression applied on large dataset. Why standardize features?", "options": [ "Data shrunk automatically", "Alpha irrelevant", "Intercept removed otherwise", "L1 penalty treats all features fairly only if scaled" ], "correctAnswerIndex": 3, "explanation": "Without standardization, large-scale features dominate the penalty." }, { "id": 24, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Action?", "options": [ "Remove alpha", "Ignore missing values", "Impute missing values first", "L1 handles missing automatically" ], "correctAnswerIndex": 2, "explanation": "Lasso cannot handle missing values; imputation or removal is required." }, { "id": 25, "questionText": "Scenario: Lasso Regression applied with high alpha. Many coefficients zero. Risk?", "options": [ "Intercept ignored", "Underfitting due to excessive sparsity", "Overfitting", "Training error minimized" ], "correctAnswerIndex": 1, "explanation": "Too high alpha increases bias and underfits the data." }, { "id": 26, "questionText": "Scenario: Lasso Regression applied to dataset with 100 features, many irrelevant. Observation: only 20 features have non-zero coefficients. Reason?", "options": [ "Alpha too small", "Training error minimized", "Intercept removed", "L1 penalty induces sparsity, zeroing irrelevant features" ], "correctAnswerIndex": 3, "explanation": "Lasso’s L1 penalty shrinks some coefficients to exactly zero, effectively performing feature selection." }, { "id": 27, "questionText": "Scenario: Lasso Regression applied to highly correlated features. Observation: model selects some features and zeros others. Advantage?", "options": [ "Intercept removed automatically", "All features retained", "Reduces multicollinearity impact and produces sparse model", "Increases variance" ], "correctAnswerIndex": 2, "explanation": "Lasso selects one feature from a correlated group, reducing multicollinearity and producing a simpler model." }, { "id": 28, "questionText": "Scenario: Lasso Regression applied with cross-validation on dataset with noisy features. Best practice?", "options": [ "Always use alpha=1", "Select alpha minimizing validation/test error", "Remove polynomial terms", "Ignore feature scaling" ], "correctAnswerIndex": 1, "explanation": "Cross-validation finds alpha that balances bias and variance, improving generalization on noisy data." }, { "id": 29, "questionText": "Scenario: Lasso Regression applied to polynomial regression of degree 8. Observation: some high-degree term coefficients zeroed. Why?", "options": [ "Intercept removed", "Training error minimized", "Polynomial terms ignored automatically", "L1 penalty shrinks less important coefficients to zero, controlling overfitting" ], "correctAnswerIndex": 3, "explanation": "Lasso penalizes large coefficients, eliminating less important high-degree terms to prevent overfitting." }, { "id": 30, "questionText": "Scenario: Lasso Regression applied to dataset with one-hot encoded categorical features. Observation: some dummy variables zeroed. Effect?", "options": [ "Training error increases", "All features retained", "Automatic feature selection among categorical levels", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Lasso can zero some dummy variables, selecting the most predictive categories." }, { "id": 31, "questionText": "Scenario: Lasso Regression applied to dataset with features in different units. Observation: large coefficients for small-scale features. Cause?", "options": [ "Data uncorrelated", "Intercept removed", "Alpha too low", "L1 penalty is unfair without standardization" ], "correctAnswerIndex": 3, "explanation": "Without standardization, features with small scales are penalized less, leading to larger coefficients." }, { "id": 32, "questionText": "Scenario: Lasso Regression applied to time-series lag features. Standardization importance?", "options": [ "Intercept ignored", "Training error minimized automatically", "Alpha irrelevant", "Ensures fair L1 penalty across features" ], "correctAnswerIndex": 3, "explanation": "Standardization ensures all features contribute fairly to the penalty." }, { "id": 33, "questionText": "Scenario: Lasso Regression applied with alpha too small. Observation?", "options": [ "Fails to converge", "Many coefficients zero", "Minimal sparsity, behaves like Linear Regression", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Small alpha provides weak L1 penalty; few or no coefficients are zeroed." }, { "id": 34, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Observation: model cannot train. Solution?", "options": [ "Increase alpha", "Ignore missing values", "Impute missing values before Lasso", "Decrease alpha" ], "correctAnswerIndex": 2, "explanation": "Lasso requires complete data; missing values must be imputed or removed before training." }, { "id": 35, "questionText": "Scenario: Lasso Regression applied to dataset with highly noisy features. Observation: fewer non-zero coefficients than Ridge. Why?", "options": [ "Alpha irrelevant", "Intercept removed", "L1 penalty zeroes some coefficients, reducing variance", "Training error minimized" ], "correctAnswerIndex": 2, "explanation": "Lasso induces sparsity by zeroing less important coefficients, effectively reducing variance." }, { "id": 36, "questionText": "Scenario: Lasso Regression applied to dataset with polynomial features of degree 10. Observation: high-degree terms zeroed. Effect?", "options": [ "Intercept removed", "Training error minimized", "Low-degree terms ignored", "Controls overfitting by eliminating less important terms" ], "correctAnswerIndex": 3, "explanation": "Lasso shrinks high-degree coefficients to zero, reducing overfitting and model complexity." }, { "id": 37, "questionText": "Scenario: Lasso Regression applied to dataset with 1000 features, alpha tuned via cross-validation. Observation: 200 features non-zero. Interpretation?", "options": [ "Intercept removed", "Model underfits", "Optimal alpha balances sparsity and generalization", "Training error high" ], "correctAnswerIndex": 2, "explanation": "Cross-validation selects alpha that keeps predictive features while zeroing irrelevant ones." }, { "id": 38, "questionText": "Scenario: Lasso Regression vs Ridge on dataset with correlated features. Observation?", "options": [ "Ridge eliminates features; Lasso does not", "Lasso selects some features and zeros others; Ridge shrinks all coefficients", "Alpha irrelevant", "Both produce sparse models" ], "correctAnswerIndex": 1, "explanation": "Lasso promotes sparsity and feature selection; Ridge keeps all correlated features." }, { "id": 39, "questionText": "Scenario: Lasso Regression applied to dataset with irrelevant features. Test error high. Recommended solution?", "options": [ "Increase alpha excessively", "Ignore irrelevant features", "Decrease alpha to zero", "Adjust alpha via cross-validation or use Elastic Net" ], "correctAnswerIndex": 3, "explanation": "Cross-validation tuning or Elastic Net can remove irrelevant features while balancing regularization." }, { "id": 40, "questionText": "Scenario: Lasso Regression applied on standardized features. Observation: correlated feature coefficients? ", "options": [ "Intercept removed", "One selected, others zeroed", "Fail to converge", "All coefficients equal" ], "correctAnswerIndex": 1, "explanation": "L1 penalty selects one correlated feature and zeros the others." }, { "id": 41, "questionText": "Scenario: Lasso Regression applied to high-dimensional dataset (features > samples). Advantage?", "options": [ "Produces sparse model, eliminating irrelevant features", "Intercept removed", "Cannot converge", "Training error minimal" ], "correctAnswerIndex": 0, "explanation": "L1 penalty shrinks unimportant coefficients to zero, useful in high-dimensional problems." }, { "id": 42, "questionText": "Scenario: Lasso Regression applied with very high alpha. Observation?", "options": [ "High variance", "Intercept ignored", "Many coefficients zeroed, possible underfitting", "Training error minimized" ], "correctAnswerIndex": 2, "explanation": "Excessive regularization increases bias, underfitting the model." }, { "id": 43, "questionText": "Scenario: Lasso Regression applied with cross-validation. Purpose?", "options": [ "Always minimize training error", "Ignore standardization", "Select alpha that minimizes validation/test error", "Remove correlated features" ], "correctAnswerIndex": 2, "explanation": "Cross-validation identifies alpha that optimally balances bias and variance." }, { "id": 44, "questionText": "Scenario: Lasso Regression applied to polynomial regression. Observation: low-degree term coefficients non-zero, high-degree zero. Effect?", "options": [ "Low-degree terms ignored", "Training error minimized", "Reduces overfitting by removing less important high-degree terms", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Lasso controls overfitting by penalizing and zeroing high-degree terms." }, { "id": 45, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Recommended practice?", "options": [ "Increase alpha", "Impute missing values before training", "Ignore missing values", "Decrease alpha" ], "correctAnswerIndex": 1, "explanation": "Lasso cannot handle missing data; it must be imputed or cleaned first." }, { "id": 46, "questionText": "Scenario: Lasso Regression applied on dataset with noisy features. Observation: fewer non-zero coefficients than Ridge. Why?", "options": [ "Intercept removed", "Training error minimized", "L1 penalty zeroes some coefficients, reducing variance", "Alpha irrelevant" ], "correctAnswerIndex": 2, "explanation": "Lasso induces sparsity, setting some coefficients to zero, which reduces variance in noisy datasets." }, { "id": 47, "questionText": "Scenario: Lasso Regression applied to dataset with one-hot encoded features. Observation: some dummy variables zeroed. Effect?", "options": [ "Intercept removed", "Feature selection among categories", "All features retained", "Training error increased" ], "correctAnswerIndex": 1, "explanation": "Lasso may zero less predictive categories, retaining only important levels." }, { "id": 48, "questionText": "Scenario: Lasso Regression applied to polynomial features of degree 12. Observation: high-degree coefficients zeroed. Effect?", "options": [ "Reduces overfitting and model complexity", "Intercept removed", "Low-degree terms ignored", "Training error minimized" ], "correctAnswerIndex": 0, "explanation": "High-degree coefficients are penalized, reducing model complexity and overfitting." }, { "id": 49, "questionText": "Scenario: Lasso Regression applied to dataset with highly correlated features. Observation: some features zeroed. Advantage?", "options": [ "Reduces multicollinearity and simplifies model", "Fails to converge", "Increases variance", "Intercept removed" ], "correctAnswerIndex": 0, "explanation": "Lasso selects one feature from correlated group, producing a simpler, less collinear model." }, { "id": 50, "questionText": "Scenario: Lasso Regression applied with small alpha. Observation?", "options": [ "Minimal feature selection; behaves like Linear Regression", "Fails to converge", "Many coefficients zeroed", "Intercept removed" ], "correctAnswerIndex": 0, "explanation": "Small alpha provides weak L1 penalty; the model retains most features with minimal sparsity." }, { "id": 51, "questionText": "Scenario: Lasso Regression applied to a medical dataset with 500 features. Many coefficients zeroed. Interpretation?", "options": [ "Training error is zero", "L1 penalty removed irrelevant features, simplifying the model", "Model underfits due to small dataset", "Intercept removed automatically" ], "correctAnswerIndex": 1, "explanation": "Lasso shrinks less important features to zero, keeping only predictive variables, which is useful in high-dimensional data." }, { "id": 52, "questionText": "Scenario: Lasso Regression applied with correlated features. Observation: only one feature from each correlated group selected. Advantage?", "options": [ "Reduces multicollinearity and produces sparse solution", "All features retained", "Intercept removed", "Model overfits" ], "correctAnswerIndex": 0, "explanation": "L1 penalty selects one feature and zeros others, reducing multicollinearity and simplifying interpretation." }, { "id": 53, "questionText": "Scenario: Lasso Regression applied with cross-validation. Optimal alpha minimizes:", "options": [ "Validation/test error", "Training error", "Intercept value", "Number of non-zero features only" ], "correctAnswerIndex": 0, "explanation": "Cross-validation selects the alpha that balances bias-variance and gives best generalization on unseen data." }, { "id": 54, "questionText": "Scenario: Lasso Regression applied on polynomial regression of degree 12. Observation: high-degree terms zeroed. Effect?", "options": [ "Low-degree terms ignored", "Reduces overfitting by eliminating less important terms", "Training error minimized", "Intercept removed" ], "correctAnswerIndex": 1, "explanation": "Lasso penalizes high-degree coefficients, zeroing unimportant terms to prevent overfitting." }, { "id": 55, "questionText": "Scenario: Lasso Regression applied to time-series dataset with lag features. Importance of feature standardization?", "options": [ "Alpha irrelevant", "Training error minimized automatically", "Ensures fair L1 penalty across all lag features", "Intercept ignored" ], "correctAnswerIndex": 2, "explanation": "Standardization ensures features on different scales are penalized fairly." }, { "id": 56, "questionText": "Scenario: Lasso Regression applied to dataset with one-hot encoded categorical features. Observation: some dummy variables zeroed. Effect?", "options": [ "Intercept removed", "Training error increased", "Automatic feature selection among categories", "All features retained" ], "correctAnswerIndex": 2, "explanation": "Lasso may zero less predictive categories, keeping only important levels." }, { "id": 57, "questionText": "Scenario: Lasso Regression applied to high-dimensional dataset (features > samples). Advantage?", "options": [ "Intercept removed", "Cannot converge", "Produces sparse model, eliminating irrelevant features", "Training error minimal" ], "correctAnswerIndex": 2, "explanation": "L1 penalty zeros unimportant coefficients, which is effective in high-dimensional settings." }, { "id": 58, "questionText": "Scenario: Lasso Regression applied with alpha too high. Observation?", "options": [ "Training error minimized", "High variance", "Many coefficients zeroed, potential underfitting", "Intercept ignored" ], "correctAnswerIndex": 2, "explanation": "Excessive regularization increases bias, underfitting the model." }, { "id": 59, "questionText": "Scenario: Lasso Regression vs Ridge Regression. Observation: Lasso zeros some coefficients, Ridge does not. Implication?", "options": [ "Both behave identically", "Ridge produces sparse model", "Lasso performs feature selection; Ridge does not", "Alpha irrelevant" ], "correctAnswerIndex": 2, "explanation": "Lasso can zero out coefficients, effectively performing feature selection; Ridge shrinks coefficients but keeps all features." }, { "id": 60, "questionText": "Scenario: Lasso Regression applied to dataset with noisy features. Observation: fewer non-zero coefficients than Ridge. Why?", "options": [ "Intercept removed", "Alpha irrelevant", "L1 penalty zeroes less important coefficients, reducing variance", "Training error minimized" ], "correctAnswerIndex": 2, "explanation": "Lasso induces sparsity and reduces variance by eliminating less predictive/noisy features." }, { "id": 61, "questionText": "Scenario: Lasso Regression applied on dataset with highly correlated features. Observation: some features zeroed. Advantage?", "options": [ "Increases variance", "Fails to converge", "Intercept removed", "Reduces multicollinearity and simplifies model" ], "correctAnswerIndex": 3, "explanation": "Lasso selects one feature from correlated group, producing a simpler, less collinear model." }, { "id": 62, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Observation: model cannot train. Recommended solution?", "options": [ "Impute missing values before training", "Increase alpha", "Decrease alpha", "Ignore missing values" ], "correctAnswerIndex": 0, "explanation": "Lasso requires complete data; missing values must be imputed or removed." }, { "id": 63, "questionText": "Scenario: Lasso Regression applied to polynomial features of degree 10. Observation: high-degree term coefficients zeroed. Effect?", "options": [ "Training error minimized", "Low-degree terms ignored", "Reduces overfitting and model complexity", "Intercept removed" ], "correctAnswerIndex": 2, "explanation": "Lasso shrinks unimportant high-degree coefficients to zero, preventing overfitting." }, { "id": 64, "questionText": "Scenario: Lasso Regression applied to a dataset with many irrelevant features. Observation: test error high. Recommended solution?", "options": [ "Increase alpha excessively", "Decrease alpha to zero", "Ignore irrelevant features", "Adjust alpha via cross-validation or use Elastic Net" ], "correctAnswerIndex": 3, "explanation": "Cross-validation tuning or Elastic Net helps remove irrelevant features and improves generalization." }, { "id": 65, "questionText": "Scenario: Lasso Regression applied to standardized features. Observation: correlated feature coefficients?", "options": [ "One selected, others zeroed", "All coefficients equal", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "L1 penalty selects one correlated feature and zeros the others." }, { "id": 66, "questionText": "Scenario: Lasso Regression applied to high-dimensional data (features > samples). Observation: model converges with sparse solution. Why?", "options": [ "L1 penalty promotes sparsity, retaining only important features", "Training error minimized", "Intercept removed", "Model underfits" ], "correctAnswerIndex": 0, "explanation": "Lasso shrinks less important coefficients to zero, allowing sparse solution even in high-dimensional datasets." }, { "id": 67, "questionText": "Scenario: Lasso Regression applied to noisy features. Observation: model zeroed several noisy features. Effect?", "options": [ "Reduces variance and improves generalization", "Intercept removed", "Training error minimized", "Alpha irrelevant" ], "correctAnswerIndex": 0, "explanation": "Lasso zeros noisy features, reducing variance and improving generalization." }, { "id": 68, "questionText": "Scenario: Lasso Regression applied to polynomial features. Observation: only low-degree coefficients retained. Reason?", "options": [ "High-degree coefficients penalized and zeroed by L1", "Training error minimized", "Intercept removed", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "Lasso penalizes high-degree coefficients, shrinking unimportant terms to zero." }, { "id": 69, "questionText": "Scenario: Lasso Regression applied to dataset with one-hot encoded categorical features. Observation: some levels zeroed. Effect?", "options": [ "Feature selection among categories", "All levels retained", "Intercept removed", "Training error increased" ], "correctAnswerIndex": 0, "explanation": "Lasso may zero less predictive categories, retaining only important levels." }, { "id": 70, "questionText": "Scenario: Lasso Regression applied with alpha too small. Observation?", "options": [ "Few coefficients zeroed; model behaves like Linear Regression", "Many coefficients zeroed", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "Small alpha provides weak L1 penalty; model retains most features with minimal sparsity." }, { "id": 71, "questionText": "Scenario: Lasso Regression applied to dataset with highly correlated features. Observation: some features zeroed. Effect?", "options": [ "Simplifies model and reduces multicollinearity", "Increases variance", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "Lasso selects one feature from correlated groups, simplifying the model and reducing multicollinearity." }, { "id": 72, "questionText": "Scenario: Lasso Regression applied with cross-validation on dataset with noisy features. Observation: alpha selected?", "options": [ "Minimizes validation/test error, balances sparsity and generalization", "Maximizes training error", "Removes correlated features automatically", "Intercept ignored" ], "correctAnswerIndex": 0, "explanation": "Cross-validation selects alpha for optimal bias-variance tradeoff." }, { "id": 73, "questionText": "Scenario: Lasso Regression applied on polynomial features. Observation: high-degree coefficients zeroed. Reason?", "options": [ "L1 penalty shrinks less important terms to zero", "Intercept removed", "Training error minimized", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "Lasso controls overfitting by penalizing and zeroing high-degree terms." }, { "id": 74, "questionText": "Scenario: Lasso Regression applied to high-dimensional dataset. Observation: sparse solution. Advantage?", "options": [ "Improves interpretability and reduces overfitting", "Training error minimized", "Intercept removed", "Model underfits always" ], "correctAnswerIndex": 0, "explanation": "Sparse solution keeps only important features, enhancing interpretability and generalization." }, { "id": 75, "questionText": "Scenario: Lasso Regression applied with standardized features. Observation: correlated features handled by sparsity. Benefit?", "options": [ "Simpler model with reduced multicollinearity", "Training error minimized", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "Lasso selects only one feature from correlated groups, simplifying the model." }, { "id": 76, "questionText": "Scenario: Lasso Regression applied to a genomics dataset with 20,000 features and 500 samples. Observation: only 150 features non-zero. Benefit?", "options": [ "Reduces dimensionality and improves interpretability", "Training error minimized", "Intercept removed", "All features retained" ], "correctAnswerIndex": 0, "explanation": "Lasso’s sparsity selects only the most predictive genes, reducing dimensionality and simplifying interpretation." }, { "id": 77, "questionText": "Scenario: Lasso Regression applied on dataset with highly correlated financial indicators. Observation: only one indicator from each group retained. Advantage?", "options": [ "Reduces multicollinearity and simplifies model", "Increases variance", "All features retained", "Intercept removed" ], "correctAnswerIndex": 0, "explanation": "L1 penalty selects one feature from correlated groups, producing a simpler, less collinear model." }, { "id": 78, "questionText": "Scenario: Lasso Regression applied with cross-validation. Observation: optimal alpha very high. Effect?", "options": [ "Many coefficients zeroed, risk of underfitting", "All coefficients retained", "Intercept removed", "Training error minimized" ], "correctAnswerIndex": 0, "explanation": "High alpha increases regularization, shrinking many coefficients to zero and potentially underfitting the model." }, { "id": 79, "questionText": "Scenario: Lasso Regression applied to polynomial regression (degree 15). Observation: high-degree coefficients zeroed. Benefit?", "options": [ "Controls overfitting and reduces model complexity", "Training error minimized", "Intercept removed", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "Lasso penalizes high-degree terms, shrinking less important coefficients to zero and preventing overfitting." }, { "id": 80, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Action required?", "options": [ "Impute missing values before training", "Ignore missing values", "Decrease alpha", "Increase alpha" ], "correctAnswerIndex": 0, "explanation": "Lasso cannot handle missing data; it must be imputed or removed first." }, { "id": 81, "questionText": "Scenario: Lasso Regression applied on standardized one-hot encoded categorical features. Observation: some dummy variables zeroed. Effect?", "options": [ "Feature selection among categories", "All dummy variables retained", "Intercept removed", "Training error increased" ], "correctAnswerIndex": 0, "explanation": "Lasso may zero less predictive categories, retaining only important levels." }, { "id": 82, "questionText": "Scenario: Lasso Regression applied to a noisy dataset. Observation: fewer non-zero coefficients than Ridge. Why?", "options": [ "L1 penalty zeroes less important coefficients, reducing variance", "Training error minimized", "Intercept removed", "Alpha irrelevant" ], "correctAnswerIndex": 0, "explanation": "Lasso induces sparsity, setting some coefficients to zero, which reduces variance in noisy datasets." }, { "id": 83, "questionText": "Scenario: Lasso Regression vs Elastic Net. When is Elastic Net preferred?", "options": [ "When correlated features need to be retained and feature selection is desired", "Always", "When features are uncorrelated", "When alpha=0" ], "correctAnswerIndex": 0, "explanation": "Elastic Net combines L1 and L2 penalties, balancing sparsity and correlated feature retention." }, { "id": 84, "questionText": "Scenario: Lasso Regression applied with very small alpha. Observation?", "options": [ "Minimal feature selection; behaves like Linear Regression", "Many coefficients zeroed", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "Small alpha provides weak regularization, retaining most coefficients and minimal sparsity." }, { "id": 85, "questionText": "Scenario: Lasso Regression applied to polynomial regression with high-degree terms dominating. Observation: high-degree terms zeroed. Benefit?", "options": [ "Reduces overfitting by removing less important terms", "Training error minimized", "Intercept removed", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "Lasso penalizes high-degree coefficients, reducing overfitting and improving generalization." }, { "id": 86, "questionText": "Scenario: Lasso Regression applied to a dataset with 10,000 features, 1,000 samples. Observation: sparse solution. Advantage?", "options": [ "Improves interpretability and reduces computational cost", "Training error minimized", "Intercept removed", "Fails to converge" ], "correctAnswerIndex": 0, "explanation": "Sparse solution retains only important features, improving interpretability and computational efficiency." }, { "id": 87, "questionText": "Scenario: Lasso Regression applied to dataset with correlated features. Observation: some coefficients zeroed. Effect?", "options": [ "Simplifies model and reduces multicollinearity", "Increases variance", "Intercept removed", "Training error minimized" ], "correctAnswerIndex": 0, "explanation": "L1 penalty selects one feature from correlated groups, producing a simpler, more stable model." }, { "id": 88, "questionText": "Scenario: Lasso Regression applied with cross-validation. Observation: alpha selected minimizes:", "options": [ "Validation/test error", "Training error", "Number of features", "Intercept value" ], "correctAnswerIndex": 0, "explanation": "Cross-validation selects alpha to balance bias-variance and achieve optimal generalization." }, { "id": 89, "questionText": "Scenario: Lasso Regression applied on standardized features with highly correlated groups. Observation: some features zeroed. Reason?", "options": [ "L1 penalty favors sparsity, selecting one feature from each correlated group", "Intercept removed", "Training error minimized", "All coefficients retained" ], "correctAnswerIndex": 0, "explanation": "Lasso promotes sparsity, zeroing less important features in correlated groups." }, { "id": 90, "questionText": "Scenario: Lasso Regression applied on dataset with polynomial features. Observation: only low-degree coefficients retained. Why?", "options": [ "High-degree coefficients penalized and zeroed", "Training error minimized", "Intercept removed", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "Lasso penalizes less important high-degree coefficients, preventing overfitting." }, { "id": 91, "questionText": "Scenario: Lasso Regression applied to dataset with missing values. Observation: model fails. Action required?", "options": [ "Impute missing values first", "Ignore missing values", "Decrease alpha", "Increase alpha" ], "correctAnswerIndex": 0, "explanation": "Lasso requires complete data; missing values must be imputed or removed." }, { "id": 92, "questionText": "Scenario: Lasso Regression applied with alpha very high. Observation: almost all coefficients zeroed. Risk?", "options": [ "Underfitting due to excessive sparsity", "Overfitting", "Intercept removed", "Training error minimized" ], "correctAnswerIndex": 0, "explanation": "Excessive alpha increases bias and underfits the model." }, { "id": 93, "questionText": "Scenario: Lasso Regression applied to standardized one-hot encoded categorical variables. Observation: some levels zeroed. Effect?", "options": [ "Simplifies categorical feature selection", "Intercept removed", "All levels retained", "Training error increased" ], "correctAnswerIndex": 0, "explanation": "Lasso zeros less predictive categories, keeping only important levels." }, { "id": 94, "questionText": "Scenario: Lasso Regression applied to dataset with 100 correlated features. Observation: only a few selected. Advantage?", "options": [ "Reduces multicollinearity and simplifies interpretation", "Increases variance", "Intercept removed", "Training error minimized" ], "correctAnswerIndex": 0, "explanation": "L1 penalty selects a subset of features from correlated groups, reducing complexity." }, { "id": 95, "questionText": "Scenario: Lasso Regression applied to polynomial features of degree 20. Observation: only low-degree terms retained. Reason?", "options": [ "High-degree terms penalized and zeroed", "Intercept removed", "Training error minimized", "Low-degree terms ignored" ], "correctAnswerIndex": 0, "explanation": "L1 penalty zeroes less important high-degree coefficients to control overfitting." }, { "id": 96, "questionText": "Scenario: Lasso Regression applied to high-dimensional sparse data. Observation: sparse solution obtained. Benefit?", "options": [ "Interpretability and reduced computational cost", "Training error minimized", "Intercept removed", "Model fails to converge" ], "correctAnswerIndex": 0, "explanation": "Sparse solution retains only relevant features, improving interpretability and efficiency." }, { "id": 97, "questionText": "Scenario: Lasso Regression applied to dataset with noisy features. Observation: model zeroed some noisy features. Effect?", "options": [ "Reduces variance and improves generalization", "Intercept removed", "Training error minimized", "All features retained" ], "correctAnswerIndex": 0, "explanation": "Lasso removes less predictive/noisy features, reducing variance and improving generalization." }, { "id": 98, "questionText": "Scenario: Lasso Regression applied with cross-validation. Observation: alpha selected reduces validation error. Advantage?", "options": [ "Balances bias-variance tradeoff and improves generalization", "Minimizes training error only", "Removes correlated features automatically", "Intercept removed" ], "correctAnswerIndex": 0, "explanation": "Cross-validation selects alpha that optimally balances bias and variance for unseen data." }, { "id": 99, "questionText": "Scenario: Lasso Regression applied to one-hot encoded dataset with missing values. Recommended action?", "options": [ "Impute missing values before training", "Ignore missing values", "Decrease alpha", "Increase alpha" ], "correctAnswerIndex": 0, "explanation": "Lasso cannot handle missing data; it must be imputed or removed first." }, { "id": 100, "questionText": "Scenario: Lasso Regression applied to dataset with high multicollinearity, noisy features, and many irrelevant variables. Best approach?", "options": [ "Standardize features, tune alpha via cross-validation, consider Elastic Net if feature selection needed", "Use Linear Regression", "Ignore alpha", "Remove L1 penalty" ], "correctAnswerIndex": 0, "explanation": "Standardization, cross-validated Lasso, or Elastic Net handles noise, multicollinearity, and feature selection effectively." } ] }