ISYE 6414 - Unit 5 || with 100% Accurate Solutions.
9 keer bekeken 0 keer verkocht
Vak
ISYE 6414 - Unit 5
Instelling
ISYE 6414 - Unit 5
What are three problems that variable selection tries to minimize? correct answers high dimensionality, multicollinearity, prediction vs explanatory
high dimensionality correct answers In linear regression, when the number of predicting variables P is large, we might get better predictions by om...
what are three problems that variable selection tr
Geschreven voor
ISYE 6414 - Unit 5
ISYE 6414 - Unit 5
Verkoper
Volgen
FullyFocus
Ontvangen beoordelingen
Voorbeeld van de inhoud
ISYE 6414 - Unit 5 || with 100% Accurate Solutions.
What are three problems that variable selection tries to minimize? correct answers high
dimensionality, multicollinearity, prediction vs explanatory
high dimensionality correct answers In linear regression, when the number of predicting
variables P is large, we might get better predictions by omitting some of the predicting variables.
Models with many predictors have... correct answers low bias, high variance
Models with few predictors have... correct answers high bias but low variance
prediction risk correct answers a measure of the bias-variance tradeoff
How do we estimate prediction risk? correct answers we can use an approach called Training
Risk
training risk correct answers compute the prediction risk for the observed data and take the sum
of squared differences between fitted values for sub model S and the observed values
Is training risk biased? why or why not? correct answers Yes, the training risk is a biased
estimate of prediction risk because we use the data twice. Once for fitting the model S and once
for estimating the prediction risk. Thus, training risk is biased upward.
The larger the number of variables is for training risk.... correct answers the larger the training
risk
What can we do since the training risk is biased? correct answers We need to correct for this bias
by penalizing the training risk by adding a complexity penalty.
Mallow's Cp correct answers This is the oldest approach to variable selection. This assumes that
we can estimate the variance from the full model, however this is NOT the case when p is larger
than n.
Akaike Information Criterion (AIC) correct answers A more general approach, for linear
regression under normality this becomes training risk + penalty that looks like Mallow's
EXCEPT the variance is the true variance not the estimate.
Leave-One-Out Cross Validation (LOOCV) correct answers This is a direct measure of
predictive power. This is just like Mallow's, except the variance is for the S submodel, not the
full model. The LOOCV penalizes complexity less than Mallow's Cp.
To correct for complexity for GLM, what can we use? correct answers AIC and BIC
, AIC vs BIC correct answers BIC is similar to AIC except that the complexity is penalized by
log(n)/2
An important aspect in prediction is.... correct answers how it performs in new settings.
We'd like to have prediction with... correct answers low uncertainty for new settings.
If p(predictors) is large, is it feasible to fit a large number of submodels? correct answers No
If p is large, what can we do instead? correct answers We can perform a heuristic search, like
stepwise regression
If p is small,.... correct answers fit all submodels
Forward stepwise regression correct answers we start with no predictor or with a minimum
model, and add one predictor at a time
Backward stepwise regression correct answers we start with all predictors, the full model and
drop one predictor at a time.
Forward-Backward stepwise regression correct answers meaning adding and discarding one
variable at a time iteratively.
Stepwise regression is a greed algorithm, what does that mean? correct answers It does not
guarantee to find the model with the best score.
Forward stepwise tends to select... correct answers smaller models
Which is preferred, forward or backward? correct answers Forward, because it selects smaller
models versus backwards which starts with a full model.
Do the three stepwise approaches select the same model? correct answers No, especially when p
is large.
Which is more computationally expensive, forward or backward? correct answers Backward
When is Mallow's Cp useful? correct answers When there are no control variables
Penalized or regularized regression correct answers When we perform variable selection and
estimation simultaneously.
If we add the bias squared and the variance, correct answers we get Mean Squared Error (MSE)
Introducing some bias yields a decrease in.... correct answers MSE
The bigger the lambda,..... correct answers the bigger the penalty for model complexity.
Voordelen van het kopen van samenvattingen bij Stuvia op een rij:
Verzekerd van kwaliteit door reviews
Stuvia-klanten hebben meer dan 700.000 samenvattingen beoordeeld. Zo weet je zeker dat je de beste documenten koopt!
Snel en makkelijk kopen
Je betaalt supersnel en eenmalig met iDeal, creditcard of Stuvia-tegoed voor de samenvatting. Zonder lidmaatschap.
Focus op de essentie
Samenvattingen worden geschreven voor en door anderen. Daarom zijn de samenvattingen altijd betrouwbaar en actueel. Zo kom je snel tot de kern!
Veelgestelde vragen
Wat krijg ik als ik dit document koop?
Je krijgt een PDF, die direct beschikbaar is na je aankoop. Het gekochte document is altijd, overal en oneindig toegankelijk via je profiel.
Tevredenheidsgarantie: hoe werkt dat?
Onze tevredenheidsgarantie zorgt ervoor dat je altijd een studiedocument vindt dat goed bij je past. Je vult een formulier in en onze klantenservice regelt de rest.
Van wie koop ik deze samenvatting?
Stuvia is een marktplaats, je koop dit document dus niet van ons, maar van verkoper FullyFocus. Stuvia faciliteert de betaling aan de verkoper.
Zit ik meteen vast aan een abonnement?
Nee, je koopt alleen deze samenvatting voor €8,32. Je zit daarna nergens aan vast.