100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached
logo-home
Samenvatting SPO - Statistische modellen 2 (Nederlands, boek: Statistiek in de praktijk) $11.29
Add to cart

Summary

Samenvatting SPO - Statistische modellen 2 (Nederlands, boek: Statistiek in de praktijk)

 79 views  1 purchase
  • Course
  • Institution
  • Book

In deze samenvatting worden alle hoofdstukken van statistische modellen 2 behandeld in het Nederlands. Daarnaast heb ik ook de korte samenvattingen toegevoegd van het Engelse boek Statistical methods for the Social Sciences. Met deze samenvatting heb ik een 9,5 gehaald, dus ik hoop dat het jou ook ...

[Show more]

Preview 4 out of 53  pages

  • No
  • 10, 11, 12, 13 en 16
  • June 29, 2021
  • 53
  • 2020/2021
  • Summary
avatar-seller
Samenvatting Statistische Modellen 2

Boek: Statistical methods for the Social Sciences
Hoofdstuk 9: Lineair Regression and Correlation (46 pagina’s)
Methods for analyzing quantitative response and explanatory variables.

Regression analysis:
1. Association between the variables → hypothesis of statistical independence
2. Strength of their association → correlation
3. Regression equation → predicts value of the response variable form the value of the
explanatory variable.
9.1 Linear Relationships
Response variable y
Explanatory variable x
Linear function The formula y = α + βx expresses observations on y as a linear function of
observations on x. The formula has a straight line graph with slope β (beta)
and y-intercept α (alpha, constante → x=0)
Origin Where x=0 and y=0
y-intercept Where x=0, so only α
Slope β The change in y for a one-unit increase in x. (x=0 → x=1)
Coördinaten (x,y)

9.7 Chapter summary
Association between quantitative variables. A regression model to describe the form of the
relationship between the explanatory variable x and the mean E(y) of the response variable. The
major aspects of the analysis are as follows:
➢ The linear regression equation E(y) = α + βx describes the form of the relationship. This
equation is appropriate when a straight line approximates the relationship between x and
the mean of y.
➢ A scatterplot views the data and check whether the relationship is approximately linear. If it
is, the least squars estimates of the y-intercept α and the slope β provide the prediction
equation ^y = a + bx closest to the data in terms of a sum of squared residuals.
➢ The correlation r and its square describe the strength of the linear association. The
correlation is a standardized slope, having the same sighn as the slope but falling between -1
and +1. Its square r2, gives the proportional reduction in variability about the prediction
equation compared to the variability about ȳ.
➢ For inference about the relationship, a t test using the slope or correlation tests the null
hypothesis of independence, namely, that the population slope and correlation equal 0. A
confidence interval for the slope estimates the size of the effect.

Summary of tests of independence ans measures of association
Measurement Levels of variables
Nominal Ordinal Interval
Null hypothesis H0: Independence H0: Independence H0: Independence (β = 0)
2 2
Test statistic χ = Som(f0 – fe) / fe z = ^y/SE t = b/se, df = n-2
Measure of 𝜋̂2 − 𝜋̂1 ^y = C – D / C + D r = b(sx / sy)
association
Odds ratio

, Hoofdstuk 11: (48 pagina’s)
11.9 Chapter Summary
This chapter generalized the bivariate regression model to include additional explanatory variables.
The multiple regression equation relating a response variable y to a set of k explanatory variables is:
E(y) = α + β1x1 + β2x2 + … βkxk
➢ The βi are partial regression coefficients. The value βi is the change in the mean of y for a
one-unit change in xi, controlling for the other variables in the model.
➢ The multiple correlation R describes the association between y and the collective set of
explanatory variables. It equals the correlation between the observed and predicted y-
values. It falls between 0 and 1.
➢ R2 = (TSS – SSE) / TSS represents the proportional reduction in error from predicting y using
the prediction equation ^y = a + b1x1 + b2x2 + … + bkxk instead of ȳ. It equals the square of the
multiple correlation.
➢ A partial correlation, such as ryx1*x2, describes the association between two variables,
controlling for others. It falls between -1 and +1.
➢ The squared partial correlation between y and xi represents the proportion of the variation in
y than can be explained by xi, out of that part left unexplained by a set of control variables.
➢ An F statistic tests H0: β1 = β2 = … = βk = 0, that the response variable is independent of all the
predictors. A small P-value suggests that at least one predictor affects the response.
➢ Individual t tests and confidence intervals for {βi} analyze partial effects of each predictor,
controlling for the other variables in the model.
➢ Interaction between x1 and x2 in their effects on y means that the effect of either predictor
changes as the value of the other predictors changes. We can allow this by introducing cross-
products of explanatory variables to the model, such as the term β3(x1x2).
➢ To compare regression models, a complete model and a simpler reduced model, the F-test
compares the SSE values or R2-values.
➢ Standardized regression coefficients do not depend on the units of measurement. The
estimates standardized coefficient b*I describes the change in y, in y standard deviation
units, for a one-standard-deviation increase in xi, controlling for the other explanatory
variables.

Bivariate regression Multiple regression
Model Prediction E(y) = α + βx E(y) = α + β1x1 + β2x2 + … βkxk
equation ^y = a + bx ^y = a + b1x1 + b2x2 + … + bkxk
Simultaneous effect of x1 …, xk Partial effect of one xi
b = Slope
bi = partial slope
r = correlation,
R = Multiple correlation, 0 ≤ R ≤ b*I = standardizes
standardized slope,
Properties of 1 regression coefficient
-1 ≤ r ≤ 1 2
measures R = PRE measure, Partial correlation,
r has the same sign as b
0 ≤ R2+ ≤ 1 -1 ≤ ryx1*x2 ≤ 1,
r2 = PRE measure,
Same sign as bi and b*I,
0 ≤ r2 ≤ 1
H0: βi = 0, or H0: popul,
partial corr. = 0.
Test of no H0: β = 0 or H0: ρ = 0, H0: β1 = …. = βk = 0
y not associated with xi,
association y not associated with x (y not associated with x1, …, xk)
controlling for other x
variables
F = Regression MS / Residual MS
t = b/SE = r / wortel((1-
= (R2/k)/(1-R2)/[n-(k+1)] t = bi / SE
Test statistic r2)/n-2)
Df1 = k df = n – (k + 1)
df = n-2
Df2 = n – (k + 1)

,Hoofdstuk 12 (44 pagina’s)
12.9 Chapter Summary
Anlaysis of variance (ANOVA) methods for comparing several groups according to their means on a
quantitative response variable. The groups are categories of one or more categorical explanatory
variables.
➢ One-way ANOVA methods compare means for categories of a single explanatory variable.
➢ Two-way ANOVA methods compare means across categories for each of two explanatory
variables. Assuming no interaction, the mean effects describe the effect of each predictor
while controlling for the other one.
➢ Multiple comparison methods provide confidence intervals for the difference between each
pair of means, while controlling the overall error probability. The Bonferroni method does
this using an error probability for each comparison that equals the desired overall error
probability divided by the number of comparisons.
➢ Analysis-of-variance methods are special cases of multiple regression analyses. Dummy
variables in the regression model represent categories that define the groups. Each dummy
variable equals 1 for a particular category and 0 otherwise.

Ordinary ANOVA methods compare groups with independent random samples, from the groups. For
some studies, different samples instead have the same subjects, such as when an experiment
observes subjects repeatedly over time. Methods for repeated measures ANOVA result from
regression models with random effects that represent the effect of the random sample of observed
subjects. Such methods treat within-subjects effects (for repeated measurements on subjects)
differently from between-subjects effects (for independent samples of subjects).
This chapter has modelled a quantitative response variable as a function of a categorical explanatory
variables.

ANOVA tests for comparing Several Groups on a Response variable:
Element of Test One-way ANOVA Two-way ANOVA Repeated Measures
ANOVA
1. Samples Independent Independent Dependent
2. Hypotheses H0: identical means H0: identical row H0: identical means
Ha :at least two means means Ha: at least two means
not equal H0: identical col. not equal
Means
H0: no interaction
3. Test statistic F = (Between-groups MS) F = Effect MS / MS F = Effect MS / MS
/ (Within-groups MS) error error
F = distribution F = distribution F = distribution
Df1 = g – 1 Df1 = df for effect Df1 = df for effect
Df2 = N – g Df2 = df for error Df2 = df for error
4. P-value Right-tail probability Right-tail probability Right-tail probability



Hoofdstuk 13 (29 pagina’s)
Controlling for a Covariate:
➢ One-way ANOVA: compares the mean of the response variable for several groups.
➢ Two-way ANOVA: compares means while controlling for another categorical variable.
o Quantitative control variable = covariate
▪ Comparison for this type of regression = analysis of covariance

, 13.1 Comparing means and comparing regression lines
Quantitative explanatory variable by x and a categorical explanatory variable by z
➢ When categorical predictor has 2 categories → z = dummy-variable
o Several categories → set of dummy-variable

Comparing Regression Lines
Studying the effect on x on y while controlling for z is equivalent to analyzing the regression of y on x
separately for new and older homes.
➢ No interaction → true slop of the line relating expected selling price to the size of home is
the same for new and older homes
o Lines are parallel
▪ Y-intercepts also equal → regression lines coincide
➢ Interaction → if the rate of increase in selling price as a function of size of home differed for
new and existing homes, then the two regression lines would not be parallel.




Comparing Means on y, Controlling for x
To study whether the difference in mean incomes can be explained (between man and woman) by
differing experience levels of men and women, we compare mean incomes for men and women
having equal levels of experience.
➢ No interaction → lines are parallel
o The difference between mean incomes for men and women is identical for all fixed
values of x = number of years of experience.
➢ Same line → mean income is identical at each level of experience.
➢ Interaction → regression lines are not parallel
o Mean income for men is higher than the mean income for women at all experience
levels, and the difference increases as experience increases.

The benefits of buying summaries with Stuvia:

Guaranteed quality through customer reviews

Guaranteed quality through customer reviews

Stuvia customers have reviewed more than 700,000 summaries. This how you know that you are buying the best documents.

Quick and easy check-out

Quick and easy check-out

You can quickly pay through credit card or Stuvia-credit for the summaries. There is no membership needed.

Focus on what matters

Focus on what matters

Your fellow students write the study notes themselves, which is why the documents are always reliable and up-to-date. This ensures you quickly get to the core!

Frequently asked questions

What do I get when I buy this document?

You get a PDF, available immediately after your purchase. The purchased document is accessible anytime, anywhere and indefinitely through your profile.

Satisfaction guarantee: how does it work?

Our satisfaction guarantee ensures that you always find a study document that suits you well. You fill out a form, and our customer service team takes care of the rest.

Who am I buying these notes from?

Stuvia is a marketplace, so you are not buying this document from us, but from seller tesswigink1. Stuvia facilitates payment to the seller.

Will I be stuck with a subscription?

No, you only buy these notes for $11.29. You're not tied to anything after your purchase.

Can Stuvia be trusted?

4.6 stars on Google & Trustpilot (+1000 reviews)

53340 documents were sold in the last 30 days

Founded in 2010, the go-to place to buy study notes for 14 years now

Start selling
$11.29  1x  sold
  • (0)
Add to cart
Added