Comprehensive Exam Review - PowerPoint PPT Presentation

1 / 54
About This Presentation
Title:

Comprehensive Exam Review

Description:

Title: No Slide Title Author: Larry Loesch Last modified by: dcreamer Created Date: 8/24/1998 1:56:28 PM Document presentation format: On-screen Show – PowerPoint PPT presentation

Number of Views:319
Avg rating:3.0/5.0
Slides: 55
Provided by: LarryL68
Category:

less

Transcript and Presenter's Notes

Title: Comprehensive Exam Review


1
Comprehensive Exam Review
Click the LEFT mouse key ONCE to continue
2
Research and Program Evaluation Part 4
Click the LEFT mouse key ONCE to continue
3
Overview of Statistics
4
Clearly, it is not possible to present a
comprehensive review of all statistics here.
Therefore, what follows is a general overview of
major principles of statistics.
There are technical exceptions to (or
variations of) most of what is presented.
However, the information provided here is
adequate for and applicable to most of the
research in the counseling profession.
5
Parametric Statistics
Use of so-called parametric statistics is based
on assumptions including that
the data represent population characteristics
that are continuous and symmetrical.
the variable(s) has a distribution that is
essentially normal in the population.
the sample statistic provides an estimate of the
population parameter.
6
Recall that variables typically involved in
research can be divided into the categories of
discrete or continuous.
Based on this distinction, (in general) all
statistical analyses can be divided into
Analyses of Relationships among variables or
Analyses of Differences based on variables
7
In the context of this general overview, all of
the variables involved in analyses of
relationships are continuous.
Similarly, for analyses of differences, at least
one variable must be continuous and at least one
variable must be discrete.
8
Analyses of Relationships
9
The simplest (statistical) relationship involves
only two (continuous) variables.
In statistics, a relationship between two
variables is known as a correlation.
Calculation of the correlation coefficient allows
us to address the question, What do we know (or
can we predict) about Y given that we know X (or
vice versa)?
10
The correlation coefficient
is used to indicate the relationship between two
variables.
is known more formally as the Pearson
Product-Moment Correlation Coefficient.
is designated by a lower case r.
ranges in values from -1.00 through 0.00 to 1.00.
11
When r -1.00, there is a perfect negative, or
inverse, relationship between the two variables.
This means that as one variable is changing, the
associated variable is changing in the opposite
direction in a proportional manner.
When r 1.00, there is a perfect positive, or
direct, relationship between the two variables.
12
This means that as one variable is changing, the
associated variable is changing in the same
direction in a proportional manner.
When r 0.00, there is a zero-order
relationship between the two variables.
This means that change in one variable is
unrelated to change in the associated variable.
13
The question to be confronted is...
How do we know if the correlation coefficient
calculated is any good?
In general, there are two major ways to evaluate
a correlation coefficient.
One method is in regard to statistical
significance.
Statistical significance has to do with the
probability (likelihood) that a result occurred
strictly as a function of chance.
14
Evaluation based in probability is like a game of
chance.
The researcher decides whether it will be a high
stakes or a low stakes situation, de-pending
on the implications of being wrong.
The results of the decision are operationalized
in the alpha level selected for the study.
15
In the language of statistics, the alpha level
(e.g., .01 or .05), sometimes called the level of
significance, represents the (proportionate)
chance that the researcher will be wrong in
rejecting the null hypothesis.
That is, the alpha level also is the probability
of making a Type I Error.
16
In the language of statistics, the p value is
the (exact) probability of obtaining the
particular result for some statistical analysis.
Technically, the p value is compared to the alpha
level to determine statistical significance if p
is less than the alpha, the result is
statistically significant.
17
Most computer programs generate p values (i.e.,
exact probabilities) from statistical analyses.
However, most journal articles report results as
comparisons of p values to alpha levels that is,
they report, for example, p lt .05, rather than,
for example, p .0471.
18
There is at least one prominent limitation in the
evaluation of a correlation coefficient based on
statistical significance.
This limitation is related to the conditions
under which the statistical significance of the
correlation coefficient is evaluated.
The critical value is the value of the
correlation coefficient necessary for it to be
statistically significant at a given alpha level
and for a given sample size.
19
In statistics, sample size is usually expressed
in regard to degrees of freedom.
For example, the degrees of freedom for a
correlation coefficient is given by df N - 2.
For the correlation coefficient, there is an
inverse relationship between the critical values
and degrees of freedom.
As the degrees of freedom (i.e., sample size)
increase, critical values (needed for statistical
significance) decrease.
20
This means that a very small correlation
coefficient can be statistically significant if
the data are from a very large sample.
Correlation coefficients cannot be evaluated as
good or bad in an absolute sense
consideration must be given to the sample size
from which the data were derived.
21
Consider two variables A and B
Another way to evaluate a correlation coefficient
is in terms of shared variance.
By definition, if A is a variable, it has
variance (i.e., not every person receives the
same score on measure A).
22
B
Similarly, because B is a variable, it has
variance, and all (i.e., 100) of the variance of
B can be represented by a circle.
23
Of interest is how much variance variables A and
B share.
24
The percentage of shared variance is equal to
25
The term r2 is known as the coefficient of
determination.
The percentage of shared variance is how much of
the variance of variable A is common to variable
B, and vice versa.
Another way to think of it is that the percentage
of shared variance is the amount of the same
thing measured by (or reflected in) both
variables.
26
  • The good news is that the shared variance
    method as a basis for evaluating a correlation
    coefficient is not dependent upon sample size.

The bad news is that there is no way to determine
what is an acceptable level of shared variance.
Ultimately, the research consumer has to be the
judge of what is a good correlation
coefficient.
27
The Pearson Product-Moment Correlation
coefficient can be used to predict one variable
from another.
Thats helpful, but has limited application
because only two variables are involved.
28
Suppose we know of the relationships between Z
and each of several other variables.
29
In multiple correlation, one variable is
predicted from a (combined) set of other
variables.
The capital letter R is used to indicate the
relationship between the set of variables and the
variable being predicted.
The variable being predicted is known as the
criterion variable, and the variables in the
set are known as the predictor variables.
30
In computing a multiple correlation coefficient,
the most desirable situation is what is known as
the Daisy Pattern.
In the hypothetical Daisy Pattern, each predictor
has a relatively high correlation with the
criterion variable...
31
and each of the predictor variables has a
relatively low correlation with each of the other
predictor variables.
32
If achieved, a true Daisy Pattern would look
something like this.
33
The multiple correlation computational procedures
lead to a weighted combination of (some of) the
predictor variables and a specific correlation
between the weighted combination and the
criterion variable.
The same two methods used to evaluate a Pearson
Product-Moment Correlation coefficient can be
used to evaluate a multiple correlation
coefficient.
34
The methods of evaluating R include
statistical significance, although the sample
size limitation concern is less problematic if
the sample is sufficient for the computa-tions.
percentage of shared variance, where the
expression R2 x 100 represents the sum of the
intersections of the predictors with the
criterion variable.
35
A canonical correlation (Rc) represents the
relationship between a set of predictor variables
and a set of criterion variables.
A canonical correlation is usually expressed as a
lambda coefficient, often Wilks Lambda, which is
the result of the statistical computations.
36
Graphically, a canonical correlation might look
like this
37
The statistical significance of the lambda
coefficient can be readily determined.
The percentage of shared variance also can be
calculated.
However, because the lambda coefficient can have
a value greater than one, the calculation of
shared variance involves more than just squaring
the lambda coefficient.
38
The following chart summarizes the nature of the
three preceding analyses of relationships.
39
Factor analysis, a special type of analysis of
relationships among variables, is a general
family of data reduction techniques.
It is intended to reduce the redundancy in a set
of correlated variables and to represent the
variables with a smaller set of derived variables
(aka factors).
Factor analyses may be computed within either of
two contexts exploratory or confirmatory.
40
Factor analysis starts with input of the raw data.
Next, an intervariable correlation matrix is
generated from the input data.
Then, using sophisticated matrix algebra
procedures, an initial factor (loading) matrix is
derived from the correlation matrix.
41
There are three major components to the factor
loading matrix.
The first is the set of item numbers, usually
arranged in sequence and hierarchical order.
The second is the factor identifications, usually
represented by Roman numerals.
The third is the factor loadings, usually
provided as hundredths - with or without the
decimal point.
42
The result might look something like this
43
An important question is, How do we know how
many factors to retain?
In factor analysis, potentially there can be as
many factors as items.
However, usually one or some combination of three
methods is used to decide how many factors to
retain.
44
One common method is to retain factors having
eigenvalues greater than one.
Each factor has an eigenvalue, which is the sum
of the squared factor loadings for the factor.
Retaining factors having eigenvalues greater than
one also is known as applying the Kaiser
Criterion.
45
A second common method is to apply the scree test.
The scree test is a visual, intuitive method of
determining how many factors to retain by
examining the graph of the eigenvalues from the
initial factor loading matrix.
46
A third possible method is based on how much of
the total variance is to be accounted for by the
retained factors.
The total possible variance is equal to the
number of items.
Therefore, the variance percentage for any factor
is the eigenvalue divided by the total number of
items, times 100.
Factors can be retained by summing these
percentages until the desired percentage is
reached.
47
Another important question in factor analysis is,
How are the relationships among the factors to
be conceptualized?
A factor is two things
Conceptually, a factor is a representation of a
construct.
However, in regard to mathematics, a factor is a
vector in n-dimensional space.
48
Factors as constructs may be separate and
entirely distinct from one another or separate
but conceptually related to one another.
Factors as vectors reflect these possibilities by
being positioned in n-dimensional space as either
perpendicular to one another or as having an
acute angle between them.
The initial factor loading matrix is rotated to
achieve the best mathematical represen-tation and
clarity among the constructs.
49
If the factors are assumed to be distinct (i.e.,
independent) from one another, the rotation is
said to be orthogonal.
An orthogonal rotation is one in which the angles
between factors are maintained as right angles
during and after the rotation.
The most common orthogonal rotation is the
Varimax procedure.
50
If the factors are assumed to be related (i.e.,
dependent) to one another, the rotation is said
to be oblique.
An oblique rotation is one in which the angles
between factors are maintained as less than right
angles during and after the rotation.
The most common oblique rotation is the Oblimin
procedure.
51
We assign to a factor a name that encompasses
whatever is reflected in the items having their
highest factor loadings on the factor.
There are a few important things to be remembered
about factor analysis.
First, a valid factor analysis requires lots of
subjects, usually a minimum of ten times the
number of subjects as items.
52
Another important point is that even though
factor analysis is a sophisticated data analysis
technique, quite a few relatively arbitrary
decisions are made by the researcher in the
process.
Selection of the context and type of factor
analysis to be used, determination of the number
of factors to retain, and naming of the factors
are just a few of the decisions to be made.
53
And finally, just because a research study
contains a factor analysis doesnt necessarily
mean that it is good research.
The validity and appropriateness of the factor
analysis must be evaluated in order to evaluate
the worth of the research.
54
This concludes Part 4 of the presentation on
RESEARCH AND PROGRAM DEVELOPMENT
Write a Comment
User Comments (0)
About PowerShow.com