0% found this document useful (0 votes)
25 views18 pages

Experimental Design Essentials

The document discusses the principles and components of experimental research, including the importance of double-blind experiments and the factors that threaten internal validity. It outlines various types of validity, confounding variables, and control procedures necessary for effective experimental design. Additionally, it emphasizes the significance of randomization and statistical significance in ensuring reliable outcomes.

Uploaded by

samreenhina
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
25 views18 pages

Experimental Design Essentials

The document discusses the principles and components of experimental research, including the importance of double-blind experiments and the factors that threaten internal validity. It outlines various types of validity, confounding variables, and control procedures necessary for effective experimental design. Additionally, it emphasizes the significance of randomization and statistical significance in ensuring reliable outcomes.

Uploaded by

samreenhina
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd

Experiments

 Why would a double-blind experiment be used?

 Discuss how the following factors may threaten internal


validity: history, maturation, testing, instrumentation,
statistical regression, selection biases, experimental
mortality, causal time-order, diffusion or imitation of
treatments, compensation, compensatory rivalry, and
demoralization.

 Does a true experimental design address the problem of


external validity?

 Identify strengths and weaknesses of the experimental


method.
Experimental Research
Experiments

 Begin with a Hypothesis


 Modify Something in a Situation
 Compare Outcomes
 Cases or People are Termed
“Subjects”
Random Assignment
 Probability of Equal Selection

 Allows Accurate Prediction

 An Alternative to Random
Assignment is Matching
Parts of the Classic
Experiment
 Treatment or Independent Variable
 Dependent Variable
 Pretest
 Posttest
 Experimental Group
 Control Group
 Random Assignment
Variations on Experimental
Design
 Pre-experimental Design
 One-shot Case Study
 One-group Pretest-Posttest Design
 Static Group Comparison

 Quasi-Experimental and Special Designs


Types of Validity
 External Validity
 Do the results apply to the broader
population?

 Internal Validity
 Is the independent variable responsible for
the observed changes in the dependent
variable?
Confounding Variables That
Threaten Internal Validity

 Maturation
 Changes due to normal growth or
predictable changes

 History
 Changes due to an event that occurs
during the study, which might have
affects the results
Confounding Variables That
Threaten Internal Validity

 Instrumentation
 Any change in the calibration of the measuring
instrument over the course of the study

 Regression to the Mean


 Tendency for participants selected because of
extreme scores to be less extreme on a retest

 Selection
 Any factor that creates groups that are not equal at
the start of the study
Confounding Variables That
Threaten Internal Validity

 Attrition
 Loss of participants during a study; are
the participants who drop out different
from those who continue?

 Diffusion of treatment
 Changes in participants” behavior in one
condition because of information they
obtained about the procedures in other
conditions
Subject Effects
 Participants are not passive
 They try to understand the study to help them
to know what they “should do”
 This behavior termed “subject effects”
 Participants respond to subtle cues about what
is expected (termed “demand characteristics”)

 Placebo effect: treatment effect that is


due to expectations that the treatment
will work
Experimenter Effects

 Any preconceived idea of the


researcher about how the
experiment should turn out

 Compensatory effects
Types of Control
Procedures
 General control procedures (applicable to
virtually all research)

 Control over subject and experimenter


effects

 Control through the selection and


assignment of participants

 Control through specific experimental design


Principles of Experimental
Design
 Control the effects of lurking
variables on the response, most
simply by comparing two or more
treatments
 Randomize
 Replicate
Randomization
 The use of chance to divide
experimental units into groups is called
randomization.

 Comparison of effects of several


treatments is valid only when all
treatments are applied to similar groups
of experimental units.
How to randomize?
 Flip a coin or draw numbers out of a hat

 Use a random number table

 Use a statistical software package or


program
 Minitab
 www.whfreeman.com/ips
Statistical Significance
 An observed effect so large that it
would rarely occur by chance is
called statistically significant.
A few more things…

 Double-blind: neither the subjects nor


the person administering the treatment
knew which treatment any subject had
received
 Lack of realism is a major weakness of
experiments. Is it possible to duplicate
the conditions that we want?

You might also like