Conducting Scientifically-Based Research in Teaching with Technology, Part I - PowerPoint PPT Presentation

1 / 55
About This Presentation
Title:

Conducting Scientifically-Based Research in Teaching with Technology, Part I

Description:

Title: Maine 2003 Author: Rhonda Christensen Last modified by: Dana Arrowood Created Date: 11/5/2003 3:10:29 PM Document presentation format: On-screen Show – PowerPoint PPT presentation

Number of Views:98
Avg rating:3.0/5.0
Slides: 56
Provided by: RhondaChr9
Category:

less

Transcript and Presenter's Notes

Title: Conducting Scientifically-Based Research in Teaching with Technology, Part I


1
Conducting Scientifically-Based Research in
Teaching with Technology, Part I
  • SITE Annual Meeting Symposium
  • Atlanta, Georgia
  • Gerald Knezek Rhonda Christensen
  • University of North Texas
  • Charlotte Owens Dale Magoun
  • University of Louisiana at Monroe
  • March 2, 2004

2
Our History of Scientifically-Based Research
  • Foundation More than ten years of
    instrumentation development/validation
  • Research based on dissertation criteria
  • Large data sets analyzed (replication of
    findings)
  • Quantitative to tell us what is happening
    Qualitative to tell us why it is happening

3
Components for Evaluation with a Research
Agenda
  • Plan for Evaluation (when writing the grant - not
    after you get it)
  • Use reliable/valid instruments and/or
  • Work on developing instruments the first year
  • Get baseline data - how can you know how far you
    have come if you dont know where you started
  • Use comparison groups such as other PT3 grantees

4
Common Instruments
  • Stages of Adoption of Technology
  • CBAM Levels of Use
  • Technology Proficiency Self Assessment
  • Teachers Attitudes Toward Computers (TAC)

5
Online Data Acquisition System
  • Provided by UNT
  • Unix/Linux Based
  • Stores Data in Files
  • Data Shared with Contributors

6
Why are we gathering this data?
  • Campbell, D. T. Stanley, J. C. (1966).
    Experimental and Quasi-Experimental Designs for
    Research on Teaching. From Gage, N. L. (Ed.)
    Handbook of Research on Teaching. Boston Rand
    McNally, 1963.
  • Frequently references
  • McCall, W. A. (1923). How to Experiment in
    Education.

7
(No Transcript)
8
Adding Research Agendas to Evaluation
  • By experiment we refer to that portion of
    research in which variables are manipulated and
    their effects upon other variables are observed.
    (Campbell Stanley, 1963, p. 1)
  • Dependent outcome variable predicted or
    measured we hope this depends on something
  • Independent predictor variable one manipulated
    to make, or believed to make a difference
  • Did changing x influence/impact/improve y?
  • Y f(x)

9
Longitudinal Designs
  • PT3/Univ. of North Texas 1999-2003
  • Baseline data year 1
  • Pre-post course measures over multiple years
  • Trends in exit student survey data
  • PT3/University of Nevada/Reno 2003-2006
  • Best features of UNT plus comparisons w/UNT
  • Added random selection of 30-60 teachers to track
    retention through end of induction year

10
Stages of Adoption of TechnologyFall 1998
11
CECS 4100 Technology SkillsPre and Post - Spring
1999
12
What is the Experiment here?
  • Dependent variables Email, WWW, Integrated
    Applications, Teaching with Technology
    Competencies
  • Independent Variable completion of content of
    course (CECS 4100, Computers in Education)

13
Longitudinal Trends in Integration Abilities
(Research Item)
14
(No Transcript)
15
Growth in Technology Integration Course at Univ.
of North Texas (Typical PT3 Evaluation Item)
16
Data Sharing with PT3 Projects
  • Control groups are difficult
  • Comparisons within CE clusters is easy!
  • Similar trends are positive confirmations for
    each other

17
Spring 2002 Snapshot Data
  • Univ. North Texas
  • Texas AM Univ.
  • St. Thomas of Miami
  • Univ. Nevada at Reno
  • Northwestern Oklahoma State Univ.
  • Wichita State University (Kansas)

18
Demographics Spring 2002
  • 481 subjects from 5 schools for pretest
  • UNT 179
  • TAMU 65
  • Miami 14
  • Nevada 91
  • Oklahoma 95
  • Wichita St. 37
  • 157 subjects from 3 schools for post test
  • UNT, TAMU, St. Thomas (2 times)

19
Demographics Spring 2002 (cont.)
  • Age Wichita State students are older
  • Mean 28 years
  • Gender UNT TAMU have more females
  • 85 and 97
  • Graduation UNT, Nevada, Oklahoma students expect
    to graduate later
  • Teaching Level TAMU students Elem.

20
Educational Technology Preservice Courses
21
Educational Technology Preservice Courses
22
What is the Experiment here?
  • Dependent Variable Gains in technology
    integration proficiency
  • Independent Variables
  • Completion of course content (as before)
  • Comparisons/contrasts among different
    environments/curricular models (value added)

23
General Findings
  • Reliability of Stages is High
  • (r .88 test-retest)
  • Reliability of Skill Self-Efficacy Data is High
  • (Alpha .77 to .88 for 4 TPSA scales)
  • Gender Females are higher in Web Access, Home
    Computer Use, and WWW Skills

24
(No Transcript)
25
Pre-Post Trends for TAMUTwo Teacher Preparation
Courses
26
Impact Across 2 Schools (Pre-Post, UNT TAMU)
  • Stages ES .42 to .76
  • CBAM LOU ES .73 to 1.15
  • TPSA-IA ES .18 to .82
  • TPSA-TT ES .33 to 1.12
  • TPSA-WWW ES .05 to .49

27
How to Interpret Effect Size
  • Cohens d vs. other
  • Small (.2), medium (.5) vs. large (.8)
  • Compare to other common effect sizes
  • As a quick rule of thumb, an effect size of 0.30
    or greater is considered to be important in
    studies of educational programs. (NCREL)
  • For example .1 is one month learning (NCREL)
  • others

SRI International. http//www.ncrel.org/tech/claim
s/measure.html
28
APA Guidelines for Effect Size
  • The Publication Manual of the American
    Psychological Association (APA, 2001) strongly
    suggests that effect size statistics be reported
    in addition to the usual statistical tests. To
    quote from this venerable guide, "For the reader
    to fully understand the importance of your
    findings, it is almost always necessary to
    include Some index of effect size or strength of
    relationship in your Results section" (APA, 2001,
    p. 25). This certainly sounds like reasonable
    advice, but authors have been reluctant to follow
    this advice and include the suggested effect
    sizes in their submissions. So, following the
    lead of several other journals, effect size
    statistics are now required for the primary
    findings presented in a manuscript.

29
UNR Collaborative Exchange
30
New PT3 Project
  • Univ. of Nevada - Reno is lead and IITTL at UNT
    as outside evaluator
  • One component - following teachers after they
    graduate from the teacher ed. Program
  • Randomly select from a pool of 2004 graduates and
    contact them prior to graduation pay a stipend
    to continue in the project by providing yearly
    data

31
Procedure for Unbiased Selection
  • Locate prospective graduates to be certified to
    teach during spring 2004
  • Number consecutively
  • Use random number table to select a preservice
    candidate from the list
  • Verify student completed technology integration
    course with B or better
  • Invite preservice candidate to participate during
    induction year and possibly beyond
  • Repeat process until 60 agree to participate

32
From Edwards, A. L. (1954). Statistical Methods
for the Behavioral Sciences. NY Rinehart.
33
Maine 2003
34
(No Transcript)
35
Maine Learning Technology Initiative (MLTI)
  • 2001-2002 Laptops for all 7th graders
  • 2002-2003 Laptops for all 7th and 8th graders in
    the whole state of Maine
  • Maine Learns is About Curriculum

36
(No Transcript)
37
(No Transcript)
38
(No Transcript)
39
(No Transcript)
40
(No Transcript)
41
(No Transcript)
42
(No Transcript)
43
(No Transcript)
44
Interesting Aspects of Research
  • Sample or Population (all 17,000 students in the
    state)
  • Selection of Exploratory Schools (if wished to
    participate, one from each region)
  • Statistical measures of significance
  • Strong reliance on Effect Size

45
Research Design
  • 9 Exploration schools (1 per region)
  • Compared with 214 others
  • Used 8th grade state-wide achievement
  • Examined trend over 3 years in math, science,
    social studies, and visual/performing arts
  • Intervention -
  • Extensive teacher preparation
  • Laptop and software for every 7th-8th
    teacher/student
  • Some permitted to take home, others not

46
2003 Findings
  • Evaluators reports
  • Achievement Effect Sizes
  • Student self reports on
  • Attitudes toward school
  • Self Concept
  • Serendipitous findings are the sometimes the most
    valuable
  • Home Access
  • Gender Equity

47
(No Transcript)
48
(No Transcript)
49
Would Cohen Have Predicted This Effect?
  • "Small Effect Size d .2. In new areas of
    research inquiry, effect sizes are likely to be
    small (when they are not zero!). This is because
    the phenomena under study are typically not under
    good experimental or measurement control or both.
    When phenomena are studied which cannot be
    brought into the laboratory, the influence of
    uncontrollable extraneous variables ("noise')
    makes the size of the effect small relative to
    these (makes the 'signal' difficult to detect).
    Cohen, J. (1977), p. 25.

50
Exploratory - as Illustrated by
51
(No Transcript)
52
Contrast with Louisiana Confidence Intervals
(Teacher Perceptions of Impact)
53
Teachers' Perception of Usefulness of ARTS to the
Delta for Math and Reading vs. Fostering
Interest in Music, Learning, or Education in
General N Math Mean Math
SD Music SD t Signif 22 2.41 1.05 3.09 2.41 1.87
0.068 not quite significant 22 Postive
Learning Experience 22 2.41 1.05 3.05 1.33 1.7
7 0.0837 not quite significant 22 Positive
Effect on Education 2.41 1.05 2.95 1.4 1.45 0
.1552 not statistically significant 22
Reading Reading SD Music SD t Signif 22 2.32 1
3.09 1.34 2.16 0.0365 statistically
significant 22 Postive Learning
Experience 22 2.32 1 3.05 1.33 2.06 0.0459 sta
tistically significant 22 Positive Effect on
Education 22 2.32 1 2.95 1.4 1.72 0.0932 not
quite significant
54
(No Transcript)
55
Its all About Confidence
  • As shown in Figure 1, three of the measures 95
    confidence intervals are roughly 3/4 of a
    confidence interval band above that is, no more
    than 1/4 of the 95 confidence interval range
    overlaps from the upper to the lower group.
    Differences in this range are as a rule-of-thumb
    meaningful according to emerging APA
    guidelines, and roughly comparable to a p .05
    level of significance (Cumming, 2003). The effect
    size for the combined upper three versus the
    lower two is approximately ((3.093.052.95)/3)
    ((2.322.41)/2/ ((1.341.331.401.001.05)/5)
    (3.03 2.37) / 1.22 .66 / 1.22 .54,
    considerably larger than the .30 cutoff beyond
    which technology interventions are considered
    meaningful (Bialo Sivin-Kachala, 1996).
    Teachers rated the ARTS to the Delta class as
    much more useful for promoting interest in music
    and creating a positive effect on students
    overall education experience that for improving
    reading and math skills.
Write a Comment
User Comments (0)
About PowerShow.com