Political Science 585 Techniques of Political Analysis - PowerPoint PPT Presentation

1 / 36
About This Presentation
Title:

Political Science 585 Techniques of Political Analysis

Description:

In both quantitative and qualitative techniques, we proceed from the theory ... would be systematically overstated in some areas and understated in other areas. ... – PowerPoint PPT presentation

Number of Views:52
Avg rating:3.0/5.0
Slides: 37
Provided by: toddm69
Category:

less

Transcript and Presenter's Notes

Title: Political Science 585 Techniques of Political Analysis


1
Political Science 585Techniques of Political
Analysis
  • Concepts, Variables and Measurement
  • April 3, 2007

2
Key terms in todays lecture
  • Quantitative and qualitative data
  • Operationalization
  • Working hypothesis
  • Indicator
  • Dummy variable
  • Multidimensional concepts
  • Measurement error
  • Validity (face and construct)
  • Reliability

3
Quantitative and qualitative techniques
  • In both quantitative and qualitative techniques,
    we proceed from the theory/hypothesis stages by
    operationalizing the relevant concepts,
    especially our independent and dependent
    variables.
  • Most of our focus, however, will be on
    quantitative techniques. This means that no
    matter what measurement strategy we use, our
    final product will be numerical in nature.
  • For example, if we measure income precisely (i.e.
    a persons score might be 30K) the values will be
    obviously numerical.
  • However, if we choose a categorical measurement
    technique (i.e. ask people if they are rich,
    middle income or poor, we will translate
    those scores into numerical values (i.e. 2,1,0)

4
Quantitative and qualitative techniques (cont.)
  • Note that it is not the presence of verbal scores
    that makes a technique qualitative.
  • Asking a question about a persons religion
    yields verbal answers, but we translate these
    measures into numerical values.
  • On the contrary, in qualitative techniques, our
    data is mostly in the forms of prose,
    discussions, and we do not generally seek to
    translate these responses into numerical
    categories.
  • In this case, the data need not be translated
    into numbers because the data analysis does not
    use numbers. It uses argument and interpretation.
  • That having been said, most of todays discussion
    is primarily, if not exclusively, focused on an
    end goal of quantitative analysis.

5
Two major measurement decisions
  • To decide how to measure the concept captured by
    each of our key variables, we must make two key
    decisions
  • The first is substantive what concrete referent
    or referents in the real world is going to
    represent our abstract concept.
  • The second is the level and precision of
    measurement.

6
Measurement
  • In order to begin the process of testing our
    hypotheses, we need to figure out what concrete
    measures are going to stand in for our abstract
    concepts. We call this operationalization.
  • First, we should note that measurement decisions
    have little to do with whether the
    variable/concept in question is the independent
    or dependent variable.
  • It will matter later on when we choose techniques
    of data analysis, but it should not matter in
    measurement decisions.

7
Operationalizing Concepts
  • Generally, we are faced with one of two scenarios
    with respect to data
  • Sometimes, we need to take this step in order to
    create an instrument a survey question, or
    rubric for collecting data.
  • For example, we want to find out about
    individuals ideology, but dont know what
    questions to ask them.
  • Other times, we have data, but we must consider
    what abstract concepts our data can stand in for.
  • For example, we have data on the number of
    terrorist attacks, the number of deaths and
    injuries, and the costs of damages in a country,
    but we dont know which one is most appropriate
    as a measure of terrorism.

8
Example Voting
  • Lets look at a hypothesis based on our running
    example about voting
  • Theory Potential voters consider whether the
    benefits of voting outweigh the costs.
  • Hypothesis The competitiveness of an election
    will increase voting. (Note this is a different
    hypothesis than the one week looked at last time.
    As we mentioned, theories imply multiple
    hypotheses.)
  • Our variables here are competitiveness and
    voting. However, we do not yet have a sufficient
    level of concreteness.
  • Until you could tell a research associate exactly
    what data to collect and what it looks like, the
    process of operationalization is not complete.

9
Aspects of the operationalization process
  • For now, lets assume we could represent our
    concept with one measure. We will return to cases
    where we need multiple measures later.
  • The important steps in the process are
  • First, we want to analyze our concept. What is
    its essence? A good first step is always to
    define the term completely. Then we can start
    considering potential measures.
  • We want to be certain that our measure actually
    represents the concept itself, and not a related
    concept.
  • For example, ideology and party identification
    are closely related, but asking a persons party
    in order to determine their ideology is
    incorrect. The definition of ideology relates to
    principles and policy preferences, and ideology
    would exist in a country with no parties.

10
Competitiveness
  • Lets consider the concept in the voting example.
    Competitiveness is too abstract a concept in its
    current form.
  • At its essence, the concept of competitiveness is
    about whether each candidate has an equal chance
    of winning. This is the idea were trying to get
    at.
  • We might still ask a number of questions before
    choosing an indicator
  • Do we care about real competitiveness or
    perceived competitiveness?
  • Do we care about Election Day as a snapshot, or
    the whole campaign period?
  • We might settle on recent polls, actual election
    results, or individuals perceptions. At this
    point, we have what Manheim calls a working
    hypothesis.

11
Aspects of the operationalization process (cont.)
  • The process is not complete yet. Second, we want
    to consider what type of data the measure we have
    chosen will yield.
  • Lets say we decide to use opinion poll results.
    The difference between Candidates A and B, one
    week before the election captures our concept
    well. Thus, this number (As percentage minus Bs
    percentage) would be what Manheim calls our
    indicator.
  • Now, what else should we know?
  • The level of measurement for our indicator. In
    this case, the indicator is interval level.
  • The range of our indicator. In this case, (50-50)
    0 yields the lowest value (100-0) 100 yields
    the highest value.
  • The meaning of measure values. 0 indicates a very
    competitive race 100 indicates a very
    uncompetitive race.

12
Aspects of the operationalization process (cont.)
  • Finally, do we want to change the level of
    measurement? Although we generally prefer the
    highest level of measurement possible, do we have
    a compelling reason to break our measure down
    into ordered categories?
  • Lets say we want to break the above polling data
    into three categories close, competitive, and
    uncompetitive races.
  • We then must decide what values to assign to each
    category.
  • For example, we might say that races where (A-B)
    is less than 5 are close, races where (A-B) is
    between 5 and 15 are competitive and other
    races are uncompetitive.
  • These categorization decisions are a largely
    subjective part of the operationalization
    process.

13
Voting (cont.)
  • If we turn to our dependent variable, voting, we
    want to measure whether or not an individual
    voted, i.e. a yes/no question.
  • Indicators that answer a yes/no question are a
    special class of nominal level variables, called
    a dummy variable. A dummy variable takes on the
    value of 1 if the answer is yes, and 0 if the
    answer is no.
  • As you can see, this answers the questions about
    level of measurement, range, and meaning of
    values in one felt swoop.

14
One final caveat
  • Voting might seem to already be concrete and
    suggest an obvious indicator. We could count the
    number of people (or the percent) who voted in
    each race and this is our value for the dependent
    variable for that observation.
  • However, this is not in fact a good
    operationalization because our hypothesis was
    about individual behavior and our measure is of
    aggregate behavior. We want each observation to
    be a single voter.
  • Our measurement is now complete. For each
    individual, we would have generated data that
    might look like the following

15
Example what the data actually looks like
16
Choosing measures
  • Sometimes, measures are chosen due to practical
    considerations. You might have data available
    which does not perfectly capture the concept, but
    you cannot afford to collect perfect data.
  • Other times, choosing a measure requires an
    intellectual trade-off how close can you get to
    the abstract idea without the concrete version
    being too complex, unobservable, or not
    measurable?
  • For example, we might like to know more than
    whether a person voted we could measure how
    enthused they were about voting from measuring
    brain activity. Not very practical though.
  • However, often, complex theories require more
    complex operationalizations.

17
Multidimensional concepts
  • Some concepts have multiple aspects or dimensions
    at the abstract level. (Arguably, most or all
    concepts do).
  • For example, democracy might entail
  • Regular, free and fair elections
  • Multiple political parties
  • Peaceful transitions in power, etc.
  • Thus, at the concrete level, we seemingly have no
    choice but to collect multiple measures.

18
Multiple measures
  • If we are analyzing second-hand data, we may use
    multiple measures because we lack a single
    strong, conceptually faithful measure.
  • For example, if we are trying to measure ideology
    but lack an appropriate question, we might look
    at a battery of questions that ask people about
    policy positions. More questions may allow us to
    triangulate on the concept of ideology itself.

19
Scales
  • Although we may collect multiple indicators, we
    still may choose to use only one variable to
    represent the concept.
  • For example, the well-known POLITY scores
    summarize the various aspects of democracy into a
    single score.
  • Each characteristic of democracy that a country
    possesses might earn it one point, and those
    points would then be summed (or weighted, then
    summed).
  • This sum would then be the value of the variable
    democracy for that observation.
  • Chapter 9 of Manheim (not required) talks more
    about scaling techniques.

20
Working hypotheses
  • Remember that refutation of a working hypothesis
    is not exactly the same as refutation of a
    hypothesis
  • If the working hypothesis is disconfirmed, we may
    always question the operationalization of the
    variables
  • Only if we accept the working hypotheses would we
    accept that the hypotheses are disconfirmed by
    them.

21
Krasno and Green
  • The theory in this article involves the relative
    importance of national and local forces in
    convincing candidates to run.
  • There are basically two main hypotheses
  • The more positive the national forces are for a
    candidates party, the higher quality challenger
    will emerge.
  • The more positive the local forces are for a
    candidates party, the higher quality challenger
    will emerge.
  • The article is also interested in the relative
    importance of these factors. At the end of this
    class, well examine techniques for comparing two
    causal factors.
  • The variable in question is the dependent
    variable, challenger quality. The two IVs are
    relatively uncontroversial in terms of
    measurement.

22
Krasno and Green the measurement problem
  • In the literature on legislative elections,
    incumbents are difficult to defeat yet, it is
    easier to predict the success of challengers if
    we know their quality.
  • In the abstract, quality means a lot of things
    charisma, ability to raise money, ability to
    debate, experience, etc.
  • Most of these things are difficult to measure

23
Krasno and Green (cont.)
  • In the past (and indeed, still now) most studies
    use a single proxy for quality prior
    office-holding.
  • Krasno and Green suggest that we can get closer
    to the concept of quality by collecting
    additional data famous names, specific office
    held, etc.
  • Of course, collecting this data is costly and
    difficult. This finding may be worthwhile without
    being all that practicable.

24
Measurement problems
  • Although we can often justify our measurement
    decisions through argument, there are more formal
    criteria for assessing how good ones measures
    are. Most importantly, our goal is to minimize
    measurement error.
  • Measurement error is any difference between the
    recorded value and the true value.
  • Error can be systematic or random
  • If the measure creates systematic error, it has a
    validity problem
  • If the measure creates random error, it has a
    reliability problem

25
Validity
  • Validity Do our measures accurately capture the
    concepts in our theory?
  • Obviously, this first and foremost requires
    proper operationalization.
  • It also, however, requires that there are no
    systematic measurement problems. For example,
    measuring social class based on home ownership
    would ignore the reality that home ownership is
    easier, even for the working class, in some parts
    of the country. Peoples social class would be
    systematically overstated in some areas and
    understated in other areas.

26
Reliability
  • Reliability Do our measures provide stable
    measurement of the concept?
  • If a measure is not reliable, it cannot be valid
    either. The reverse does not hold.
  • Random error can come from a variety of sources
    (for a fairly complete list, see pp. 74-75 of
    Manheim)
  • Important examples include different
    interpretation of questions, recording of data,
    and different settings and temporal contexts.

27
Validity and reliability (cont.)
  • How do we assess reliability?
  • Mostly objective criteria
  • Test/retest methods. An example would be a scale.
    A reliable scale would weigh a person the same on
    two consecutive occasions.
  • Sub-sample methods are more statistically complex
    and rely on large samples. The basic idea is that
    if we cut the sample randomly in half, there will
    be no major differences between the two halves on
    the measure in question.
  • Reliability can only be estimated, not calculated.

28
Validity and reliability (cont.)
  • How to assess validity?
  • Criteria are somewhat more subjective, but
    nonetheless involve statistical analysis.
  • Please note my terms here differ slightly from
    the book, to avoid double usage of the terms
    internal and external validity (which well
    discuss next time).
  • Two ways of assessing validity
  • Face validity Is it self-evident that the
    measure represents the concept well? Put another
    way, would your fellow scholars and experts
    accept your measure?
  • Construct validity Is the measure correlated
    with other measures that the concept is related
    to?

29
Construct validity an example
  • Lets say you are asking individuals a question
    about their ideology, and you want to assess its
    validity. In other words, is this question doing
    a good job of identifying conservatives as
    conservatives, and liberals as liberals?
  • To measure construct validity, we might also ask
    the individuals who they have voted for in recent
    elections. If theres not at least a modest
    correlation between ideology and vote choice, the
    ideology measure probably has a validity problem,
    even if it makes sense conceptually.

30
Validity and reliability a few final comments
  • IMPORTANT The two types of validity on the
    previous slides refer to measures.
  • Later, we will discuss internal and external
    validity, which are criteria for research
    designs.
  • Even if a single measure has less than perfect
    validity or reliability, multiple measures can
    salvage the research design.
  • Also, validity problems may be mitigated
    depending on the scope of our study. If our study
    only covered a single city, for example, the
    above use of home ownership as a proxy for social
    class might be less problematic.

31
Krasno and Green
  • So how do Krasno and Green do on these
    assessments?
  • Examining the elements of the challenger quality
    scale, each characteristic appears to have face
    validity.
  • We might also say that the weighting decision has
    face validity because the authors justify the
    4-point award for office-holding in terms of the
    four elements it implies political contacts,
    name recognition, candidate skills, establishing
    occupational qualification.

32
Krasno and Green (cont.)
  • In addition, the measures construct validity is
    strong based on the examination of year-by-year
    scores
  • 1974 was a horrible year for Republicans
    (Watergate) and this was the Democrats best year
    in terms of relative candidate quality scores.
  • 1980 was a bad year for Democrats and likewise,
    Republicans had their best candidate quality
    scores that year.

33
Krasno and Green (cont.)
  • The reliability of this measure is not
    necessarily high because it depends on the
    authors judgments
  • What is a celebrity?
  • Where do you draw the cutoff for important
    offices?
  • What qualifies a person as professional?
  • If you or I took the same raw data and created
    the same scores, we might get different results
  • For measures such as this, coding should always
    use multiple coders and researchers should
    measure intercoder reliability. This offsets the
    problems associated with subjectivity. Well talk
    more about this later in the quarter when we
    discuss expert surveys.

34
Class exercise
  • Using the concept assigned to your group, you are
    going to choose one hypothesis in which your
    concept is the dependent variable. It may be a
    hypothesis from last time, or one you create
    today.
  • Follow the operationalization process outlined
    today, including
  • A working definition of the two key concepts and
    a discussion of its important features (concept
    analysis). Most importantly, is it a
    multidimensional concept?

35
Class exercise
  • Brainstorm some possible measures for each
    concept. Can you use one measure, or are multiple
    measures preferable? Which one are you going to
    choose?
  • What will the data generated by this measure (or
    measures) look like, in terms of level of
    measurement, range, and indicator values?
  • As before, please turn in a written copy of your
    groups notes.

36
Next time
  • Next class, we will lay out a framework for the
    examination of research techniques. This entails
    a number of considerations
  • How do we choose cases, and what implications do
    those decisions have?
  • How do we know that we can trust our results?
    (internal validity)
  • When can we generalize from our study to all
    instances of the phenomenon? (external validity)
  • What ethical considerations come into play?
  • As you read Geddes, think not only about the
    questions she poses, but also about the theory,
    hypothesis, and the measurement decisions made in
    her study.
Write a Comment
User Comments (0)
About PowerShow.com