Chapter 10: Analyzing Experimental Data - PowerPoint PPT Presentation

1 / 34
About This Presentation
Title:

Chapter 10: Analyzing Experimental Data

Description:

Conduct a statistical test to determine if the group difference or main effect is significant. ... groups means differed more than expected based on error variance ... – PowerPoint PPT presentation

Number of Views:26
Avg rating:3.0/5.0
Slides: 35
Provided by: faxse
Category:

less

Transcript and Presenter's Notes

Title: Chapter 10: Analyzing Experimental Data


1
Chapter 10 Analyzing Experimental Data
  • Inferential statistics are used to determine
    whether the independent variable had an effect on
    the dependent variance.
  • Conduct a statistical test to determine if the
    group difference or main effect is significant.
  • If we have different group means, this suggests
    that the independent variable had an effect.

2
Memory for Words and Nonwords
3
  • BUT, there can be differences between the group
    means even if the independent variable does not
    have an effect
  • The means will almost never be exactly the same
    in different conditions
  • Error variance (random variation) in the data
    will likely cause the means to differ slightly
  • So the difference between the two groups must be
    bigger than we would expect based on error
    variance alone to conclude it is significant.
  • We use inferential statistics to estimate how
    much the means would differ due to error variance
    and then test to see whether the difference
    between the two means is larger than expected
    based on error variance.

4
  • Hypothesis testing
  • Null Hypothesis (H0) states that the independent
    variable did not have an effect.
  • The data do not differ from what we would expect
    on the basis of chance or error variance
  • Experimental Hypothesis (H1) states that the
    independent variable did have an effect.
  • If the researcher concludes that the independent
    variable did have an effect they reject the null
    hypothesis.
  • groups means differed more than expected based on
    error variance

5
  • If they conclude that the independent variable
    did not have an effect they fail to reject the
    null hypothesis
  • Groups means did not differ more than expected
    based on error variance.
  • Type I error when you reject the null hypothesis
    when is in fact true.
  • The researchers conclude that the independent
    variable had an effect, when in reality it did
    not have an effect.
  • The probability of making a type 1 error is equal
    to alpha (?).

6
  • Most researchers set ? .05 meaning that they
    will make a type I error not more than 5 times
    out of 100.
  • There is a 95 probability they will correctly
    conclude there is a difference and a 5
    probability they will conclude there is a
    difference when there was not a real difference.
  • If you set ? .01 ( more conservative). You know
    that only 1 out of 100 times would expect to find
    a difference when there really is no difference.
  • 99 confident your results are do to a real
    difference and not chance or error variance.

7
  • Type II error fail to reject the null hypothesis
    when the null hypothesis is really false.
  • The researcher concludes that the independent
    variable did not have an effect when it fact it
    did have an effect.
  • The probability of making a type II error is
    equal to beta (?).
  • Many factors can result in a type II error
    unreliable measures, mistakes in data collecting,
    coding, and analyzing, a small sample size, very
    high error variance.

8
  • Power of a test is the probability that the
    researchers will be able to reject the null
    hypothesis if the null hypothesis is false.
  • The ability of the researchers to detect a
    difference if there is a difference
  • Power 1 - ?
  • Type II errors are more common when the power is
    low.

9
Researchers Decision
10
  • Power is related to the number of participants in
    a study, the greater the number of participants
    the higher the power.
  • Researchers may conduct a power analysis to
    determine the number of participants they would
    need to detect a difference.
  • Power of .80 higher is considered good (80
    chance of detecting an effect if there is an
    effect).
  • If the power is .80 then beta is .20.
  • Alpha is usually set at .05, but beta at .20,
    because it is worse to make a Type I error
    (saying there is a difference when there really
    is not) than a type II error (fail to find a
    difference when there really is a difference)

11
  • Effect Size index of the strength of the
    relation between the independent variable and the
    dependent variable.
  • The proportion of variability in the dependent
    variable that is due to the independent variable
  • ranges from .00 to 1.00
  • If the effect size is .39, this means that 39 of
    the variability in the dependent variable is due
    to the independent variable.

12
  • T test
  • An inferential test used to compare to means
  • Step 1 calculate the mean of the two groups
  • Step 2 calculate the standard error of the
    difference between the two means
  • how much the means are expected to differ based
    on error variance
  • 2a calculate the variance of each group
  • 2b calculate the pooled standard deviation
  • Step 3 Calculate t
  • Step 4 Find the critical value of t
  • Step 5 Compare calculated t to critical t to
    determine whether you should reject the null
    hypothesis.

13
  • Paired t-test is used when you have a within
    subjects design or matched subjects design. The
    participants in the the condition are either the
    same (within) or very similar (matched).
  • This test takes into account the similarity in
    the participants
  • More powerful test because the pooled variance is
    smaller resulting in a larger t.
  • Computer analyses are now used to conduct most
    tests.

14
Chapter 10 Analyzing Complex Designs
  • T-tests are used when you are comparing two
    means. But what if there are more than two levels
    of the independent variable or two-way design?
  • You could do separate t tests on all the means.
    But the more tests you conduct the increased
    chance of making a type I error.
  • If you made 100 comparisons, you would expect 5
    to be significant by chance alone (even if there
    is no effect).
  • If you did 20 tests you would expect about 1 to
    be significant by chance (Type I error).

15
  • Bonferroni adjustment used to adjust for the
    Type I error rate.
  • Divide the alpha level (.05) by the number of
    tests you conduct.
  • If doing 10 tests (.05/10 .005). Which means
    you must find a larger t for it to be significant
    (more conservative).
  • But this also increases your chance of making a
    Type II error (missing an effect when there
    really is one).

16
  • Analysis of Variance (ANOVA)
  • Used when comparing more than two means in a
    single factor study (one-way) or in a study with
    more than one factor (two- and three-way etc.).
  • Analyzes differences between means
    simultaneously, so Type I errors are not a
    problem.
  • Calculates the variance within each condition
    (error variance) and the variance between each
    condition.
  • If we have an effect of the independent variable
    then there should be more variance between
    conditions than within conditions.

17
  • F-test divide the between groups variance by the
    within groups variance. If larger the effect of
    the independent variable the larger the F

11
10
9
8
8
7.5
7
6
5.5
5
4
3
2
1
1
18
  • Total Sums of Squares
  • Subtract the mean from each score, square the
    difference, and then add them up.
  • SStotal is the total amount of variability in all
    the data.
  • Sum of Squares
    within Groups
  • SStotal
  • Sum of
    Squares b/w Groups

19
  • Sum of Squares Within-Groups (SSwg)
  • calculate the sum of the squares for each
    condition and then add these together.
  • This is the variability that is not due to the
    independent variance (error variance)
  • To get the average SSwg you must divide SSwg by
    the degrees of freedom (dfwg).
  • df is represented by n - k, n sample size and
    k number of groups or conditions.
  • SSwg/ n- k MSwg (mean square within-groups)

20
  • Sum of Squares Between Groups SSbg
  • calculate the grand mean (mean across all
    conditions).
  • If there is no effect of the IV all condition
    means should be close to the grand mean.
  • Subtract the grand mean from each of the
    condition means, squares these differences, and
    multiply this by the size of the group, and then
    sum across groups.
  • To get an average of the SSbg you must SSbg/ dfbg
  • df k - 1 (number of groups minus 1)
  • SSbg / dfbg MSbg (mean square between-groups)

21
  • This reflects the differences among the groups
    that is due to the independent variable, but
    there still may be some differences that are due
    to error variance (random variation).

22
(No Transcript)
23
  • F-test
  • Test whether the mean variance between groups is
    larger than the mean variance within groups.
  • F MSbg/ MSwg
  • If there is no effect of the independent variable
    the F value will be 1 or close to 1, the larger
    the effect the larger the F value.
  • Compare your F value to the critical F value
    using tables in text.
  • Need the alpha level (.05) and dfbg and dfwg
  • If your F is larger than the critical F then you
    can conclude that your independent variable had
    an effect or there is a significant main effect

24
  • Factorial Design
  • More than one independent variable (two-way)
  • Calculate the Mean Square (MS) for the error
    variance (within groups), independent variable A
    and B, and the interaction between A and B.
  • FA MSA/MSwg
  • FB MSB/MSwg
  • FAxB MSAxB/MSwg

25
B
1
2
1
2
A
26
(No Transcript)
27
  • Follow-Up Tests
  • If you have more than two levels of the
    independent variable the ANOVA will tell you if
    there is an effect of the independent variable,
    but it will not tell you which means differ from
    each other

28
A
B
C
29
  • Can do follow-up tests (post hocs or multiple
    comparisons) to test for differences among the
    means. Can test mean A against B and C, and B
    against C.
  • It could be that all three means differ from each
    other, or it could be that only B and C differ
    but A does not differ from B.
  • You ONLY conduct follow-up tests if the F-test
    was significant.

30
  • Interactions
  • If the interaction is significant we know that
    the effect of one independent variable differs
    depending on the level of the other independent
    variable.
  • In a 2 x 2 design (independent variable A and B)
  • Simple main effect the effect of one independent
    variable at a particular level of another
    independent variable
  • Simple main effect of A at B1, A at B2, B at A1,
    B at A2.

31
B
1
2
1
2
32
B
1
2
2
1
3
A
33
  • Multivariate Analysis of Variance (MANOVA)
  • Used when you have more than one dependent
    variable.
  • Test differences between two or more independent
    variables on two or more dependent variables
  • Why not just conduct separate ANOVAs?
  • MANOVA is usually used when the researcher has
    dependent variables that may be conceptually
    related
  • Control the Type I error rate (tests all
    dependant variables simultaneously)

34
  • MANOVA creates a new variable called the
    canonical variable (a composite variable that is
    a weighted sum of the dependent variables).
  • First, test to see if this is significant
    (multivariate F) and then conduct separate ANOVAs
    on each dependent variable.
Write a Comment
User Comments (0)
About PowerShow.com