Chalmers (2003) Trying To Do More Good Than Harm in Policy and Practice The Role of Rigorous, Transparent, Up-to-Date Evaluations
Chalmers (2003) Trying To Do More Good Than Harm in Policy and Practice The Role of Rigorous, Transparent, Up-to-Date Evaluations
Trying to Do More Good than Harm in Policy and Practice: The Role of Rigorous,
Transparent, Up-to-Date Evaluations
Author(s): Iain Chalmers
Source: Annals of the American Academy of Political and Social Science, Vol. 589, Misleading
Evidence and Evidence-Led Policy: Making Social Science More Experimental (Sep., 2003), pp. 22
-40
Published by: Sage Publications, Inc. in association with the American Academy of Political
and Social Science
Stable URL: http://www.jstor.org/stable/3658559
Accessed: 07/09/2009 02:24
Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at
http://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unless
you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you
may use content in the JSTOR archive only for your personal, non-commercial use.
Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained at
http://www.jstor.org/action/showPublisher?publisherCode=sage.
Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed
page of such transmission.
JSTOR is a not-for-profit organization founded in 1995 to build trusted digital archives for scholarship. We work with the
scholarly community to preserve their work and the materials they rely upon, and to build a common research platform that
promotes the discovery and use of these resources. For more information about JSTOR, please contact [email protected].
Sage Publications, Inc. and American Academy of Political and Social Science are collaborating with JSTOR
to digitize, preserve and extend access to Annals of the American Academy of Political and Social Science.
http://www.jstor.org
Because professionals sometimes do more harm than
good when they intervene in the lives of other people,
their policies and practices should be informed by rigor-
ous, transparent,up-to-date evaluations. Surveys often
revealwide variationsin the type and frequency of prac-
tice andpolicy interventions,and this evidence of collec-
tive uncertaintyshouldprompt the humilitythat is a pre-
condition for rigorous evaluation. Evaluation should
begin with systematicassessment of as high a proportion
as possible of existing relevant, reliable research, and
Rigorous,
Transparent, Why Do We Need Rigorous,
Up-to-Date Transparent,Up-to-Date
Evaluationsof Policy
Evaluations and Practice?
It is the business of policy makers and practi-
tioners to intervene in other people's lives.
Although they usually act with the best of inten-
By tions, however, their policies and practices
IAIN CHALMERS sometimes have unintended, unwanted effects,
and they occasionally do more harm than good.
lain Chalmers qualified in medicine in the mid-1960s
and practiced as a clinicianfor seven years in the United
Kingdom and the Gaza Strip. In the mid-1970s, after
further training at the London School of Hygiene and
TropicalMedicine and the London School of Economics
and Political Science, he became a full-time health ser-
vices researcher with a particular interest in assessing
the effects of health care. He directed the National
Perinatal Epidemiology Unit between 1978 and 1992
and the U.K. Cochrane Centre between 1992 and 2002.
DOI: 10.1177/0002716203254762
22 ANNALS, AAPSS, 589, September 2003
TRYING TO DO MORE GOOD THAN HARM 23
This reality should be their main motivation for ensuring that their prescriptions
and proscriptions for others are informed by reliable research evidence.
In her address at the opening of the Nordic Campbell Center, Merete
Konnerup, the director, gave three examples showing how the road to hell can be
paved with the best of intentions (Konnerup 2002). An analysis of more than fifty
studies suggests that effective reading instruction requires phonics and that pro-
motion of the whole-language approach by educational theorists during the 1970s
and 1980s seems likely to have compromised children's learning (National Insti-
tute of Child Health and Human Development 2000). A review of controlled
assessments of driver education programs in schools suggests that these programs
may increase road deaths involving teenagers: they prompt young people to start
driving at an earlier age but provide no evidence that they affect crash rates (Achara
et al. 2001). A review of controlled studies of "scared straight"programs for teen-
age delinquents shows that, far from reducing offending, they actually increase it
(Petrosino, Turpin-Petrosino, and Finchenauer 2000; Petrosino, Turpin-
Petrosino, and Buehler 2003).
One example of several that I could use to illustrate how my good intentions as a
medical practitioner turned out to be lethal is the advice I promulgated after read-
ing Benjamin Spock's (1966) record-breaking bestseller, Baby and Child Care. I
bought the book when I was a recent medical graduate in the mid-1960s and
marked the following passage:
ties are often manifested in a very wide range of practices, not infrequently provid-
ing indirect evidence of mutually incompatible opinions. This evidence of collec-
tive uncertainty about the effects of policies and practices should prompt
professionals and the public to find out which opinions are likely to be correct. A
lack of empirical evidence supporting opinions does not mean that all the opinions
are wrong or that, for the time being, policy and practice should not be based on
people's best guesses. On matters of public importance, however, it should prompt
efforts to obtain relevant evidence through evaluative research to help adjudicate
among conflicting opinions.
To be even more provocative,we could askwhether some of these forms of synthesis actu-
ally constitute reviewing the literature at all. A few seem to be closer to actually doing
research, ratherthan reviewing it. (Hammersley 2002, 4)
This is a remarkablytardy insight, coming as it does two decades after a fellow edu-
cational researcher published a seminal paper pointing out that "integrative
research reviews" are research projects in their own right (Cooper 1982).
Ignorance about the field of research synthesis and cavalier lack of concern
about bias in reviews may simply reflect views about the purposes of research.
Towards the end of his critique, Hammersley (2001) suggested,
Views such as this may have prompted the U.K. secretary of state for education and
employment, David Blunkett, to question the relevance of social science to govern-
ment. The decision by Blunkett's department to establish a Centre for Evidence-
Informed Policy and Practice at London's Institute of Education appears to have
been driven partly by concerns about "ideology parading as intellectual inquiry and
about the relevance and timeliness of research and the intelligibility of its results"
(Boruch and Mosteller 2002, 2).
for other purposes, is not uncommon (see, for example, Webb 2001). Researchers
need to draw on a variety of research designs (Oakley 1999, 2000; Macintyre and
Petticrew 2000), for example, to develop defining criteria for attention-deficit
hyperactivity disorder, to survey the frequency of mental illness in prison popula-
tions, to investigate the validity of methods used to assess school performance, and
to explore and record the subjective experiences of asylum seekers.
Indeed, a variety of study designs are required to assess the effects of specific
factors on some health or social characteristic, life course, or putative "outcome."
As the British sociologist John Goldthorpe (2001) has noted, a fundamental issue is
whether the researchers can manipulate the factors concerned. Often this will not
be possible, for example, in efforts to understand the effects on child development
of genetic characteristics or of divorce. Studies of the relationship between child
development and these factors may help to develop theory about the nature of the
relationship and lead to ideas about how to intervene in an effort to protect or
improve child development.
It is at this point when interventions have been conceptualized on the basis of
theory derived from observed associations-that it is important to ensure rigorous
evaluation of the effects of these interventions, for example, gene therapy, mar-
riage guidance, or child counseling. All such interventions can, in principle, be
manipulated, and empirical evaluation in controlled experiments can assess
whether they have the effects predicted by theory.
Sometimes the results of controlled experiments will be consistent with theory
and can inform the development of policy and practices. On other occasions, con-
trolled experiments will not yield evidence of the intervention effects predicted by
theory. This does not necessarily mean that the theory is wrong; but it does mean
that the possible reasons for the discrepancy between the predicted and observed
effects should be explored, possibly leading to a refinement or rejection of the the-
ory; and it should certainly be a warning that deploying the intervention in practice
may do more harm than good.
28 THE ANNALS OF THE AMERICAN ACADEMY
And a reviewer consulted by the Economic and Social Research Council about a
proposal to prepare systematic reviews of randomized trials and studies with other
designs stated (Ann Oakley, personal communication 2002),
With double bind [sic] and other safeguards generally impossible in social science
research, and typicallywith biases due to differential attrition,it is not evident that ran-
domised control trials are invariablypreferable.
Sometimes comments on randomized trials are little more than polemic and the
erection of straw men:
Randomized designs have, like all designs, important limitations. (Dobash and Dobash
2000, 257)
It is not the case, even in abstractterms, that some research designs have all the advan-
tages and others have none. (Hammersley 2001, 547).
The orthodoxy of experimental manipulation and RCTs is dangerous when applied
unthinkinglyto health promotion. (Kippax2003, 30)
30 THE ANNALS OF THE AMERICAN ACADEMY
Those who reject randomization are implying they are sufficiently knowledgeable
about the complexities of influences in the social world that they know how to take
account of all potentially confounding factors of prognostic importance, including
those they have not measured, when comparing groups to estimate intervention
effects.
Evaluationshouldbeginwith systematic
assessmentof as high a proportionas possible
of existingrelevant,reliableresearch,and then,
if appropriate,additionalresearch.
FIGURE 1
THE COCHRANE COLLABORATION?
THE COCHRANE
COLLABORATION's
result, tens of thousands of premature babies have probably suffered and died
unnecessarily (and cost the health services more than was necessary). This is just
one of many examples of the human costs resulting from failure to perform system-
atic, up-to-date reviews of randomized trials of health care.
One of the reasons that the Cochrane logo conveys the message it does is that
estimates of the effects of the treatment have been shown as 95 percent confidence
intervals. Emphasis on point estimates of effects and reliance on p values derived
from statistical tests can result in failure to detect possible effects of interventions
that may be important. This danger is illustrated in a paper by two British criminol-
ogists titled "The Controlled Trial in Institutional Research-Paradigm or Pitfall
for Penal Evaluators?' (Clarke and Cornish 1972). This drew on the authors' expe-
rience of a randomized trial of a therapeutic community for young offenders.
Because similar numbers of boys in the experimental and control groups went on to
reoffend, the authors concluded that therapeutic communities were ineffective
and that randomized trials are inappropriate for assessing the effects of institu-
tional interventions.
Had they taken account of the confidence interval surrounding the point esti-
mate of the difference between experimental and control groups, as well as the
results of other, similar studies, they might have come to a more cautious conclu-
sion (Table 1). An overall estimate of the effects of therapeutic communities based
on a systematic review of eight randomized trials suggests that this category of
intervention may halve the odds of adverse outcomes, an effect of great public
TRYING TO DO MORE GOOD THAN HARM 33
TABLE 1
SYSTEMATICREVIEW OF EIGHT RANDOMIZED CONTROLLED
TRIALS ASSESSING THE EFFECTS OF THERAPEUTIC
COMMUNITIES ON ADVERSE OUTCOMES
95 Percent
Odds Ratio Confidence Interval
All (N = 8) 0.46 0.39 to 0.54
Secure democratic
Cornish and Clarke (1975) 1.04 0.76 to 2.79
Auerbach (1978) 0.52 0.28 to 0.98
SOURCE: NHS Centre (1999).
importance if true. An analysis restricted to the two trials of the "secure demo-
cratic" model studied by Clarke and Cornish (1972) suggests that although the
beneficial effect may be somewhat less in these, the evidence is still suggestive of a
potentially very important benefit. As a consequence of a failure to take proper
account of the effects of chance, a useful methodology and a useful intervention
may both have been jettisoned prematurely.
ation. As the review progressed, the Web site showed the results of applying the
agreed inclusion and exclusion criteria and displayed the data abstracted from eli-
gible studies and eventually the draft data tables. As it happens, the investigators
were unable to identify any randomized experiments of water fluoridation, and
they were disappointed with the quality of most of the observational data
(McDonagh et al. 2000). (These suggested a modest reduction in caries and an
increase in disfiguring dental fluorosis.)
This transparent process is relevant to a point made by the president of the
Royal Statistical Society in 1996. After referring approvingly to the Cochrane Col-
laboration-which prepares, maintains, and disseminates systematic reviews of
the effects of healthcare interventions (Chalmers 1993)-he wrote,
But what's so special about medicine? We are, through the media, as ordinary citizens,
confronted daily with controversy and debate across a whole spectrum of public policy
issues. But typically, we have no access to any form of systematic "evidence base"-and
therefore no means of participating in the debate in a mature and informed manner. Obvi-
ous topical examples include education what does work in the classroom?-and penal
policy-what is effective in preventing reoffending? (Smith 1996)
It was after reading this presidential address and Robert Boruch's excellent book,
Randomized Experiments for Planning and Evaluation (1997), that I decided to
beat a path to the latter's door in October 1998. I wanted to try to persuade him to
take up the challenge of leading an effort to establish an analogue to the Cochrane
Collaboration to prepare systematic reviews of social and educational interven-
tions. For reasons that should now be clear, although I felt it was essential that such
collaboration should be international, I believed that it would fail without the lead-
ership and active involvement of social scientists in the United States, and I sug-
gested that it might be named after one of them-Donald Campbell.
The Cochrane Collaboration and the Campbell Collaboration are both exploit-
ing the advantages of electronic media. Electronic publication means that proto-
cols (containing the introduction to and materials and methods planned for each
review) as well as complete reports of systematic reviews can be made publicly
available in considerably more detail and promptly after submission than is usually
possible with print journals, and that they can be modified in the light of new data
or comments.
As far as I am aware, these two collaborations currently provide the only interna-
tional infrastructure for preparing and maintaining systematic reviews in the fields
of health and social care and education. Estimates suggest that more than ten thou-
sand people are now contributing to the Cochrane Collaboration (which was inau-
gurated in 1993), most of them through one or more of fifty Collaborative Review
Groups (all international), which have collectively published nearly two thousand
systematic reviews in The Cochrane Database of Systematic Reviews. Members of
these groups are supported by ten Cochrane Methods Groups (all international)
and twelve Cochrane Centres, which are geographically based, and share collec-
tive responsibility for global coverage (www.cochrane.org).
TRYING TO DO MORE GOOD THAN HARM 35
Our judgments can affect other people's lives, however. After comparing the
results of systematic reviews with the recommendations of experts writing text-
books and narrative review articles, Antman and his colleagues (1992) concluded
that because reviewers have not used scientific methods, advice on some life-
saving therapies has been delayed for more than a decade, while other treatments
have been recommended long after controlled research has shown them to be
harmful.
36 THE ANNALS OF THE AMERICAN ACADEMY
One needs to bear in mind Xenophanes' words and empirical evidence of this
kind when assessing nonspecific questions about the validity of systematic reviews.
Hammersley (2001, 547) is not the only person to have asked the question, "Where
is the evidence that systematic reviews produce more valid conclusions than narra-
tive reviews?"
Not only do those who pose such questions ignore the existing evidence, they
almost never confront the reality that different methods of reviewing tend to lead
to different conclusions or explore the reasons for and consequences of this. For
policy makers, practitioners, and others wishing to use research evidence to inform
ConcludingObservations
I have tried to makeclearand to justifyin this articlehow I conceptualizereli-
ableresearchevidence.Thisentailsthe preparationof systematicreviewsdesigned
to minimizebias, drawingon researchstudies designed to minimizebias. I have
deliberately concentrated on bias because the other important issue, taking
accountof the effects of chance,is a more straightforward matter(by using meta-
analysisand doing largerstudies).I believe that the principleof minimizingbias
appliesacrossallof science, andcertainlyin appliedfieldslikethe healthandsocial
sciences, because of the impactresearchmayhave on policies and practices.
In conclusion,myinterestin researchto assessthe effectsof interventionsarises
froma long-standingconcernthat,actingwiththe best of intentions,policymakers
andpractitionershave sometimesdone moreharmthangood when interferingin
the lives of others. I believe that the empiricalevidence showing associations
between studydesign and studyresults-whether amongreviewsor amongindi-
vidual studies-is likely to be explained by differentialsuccess in controlling
biases. If only as a patient, therefore, I want decisions about my care to take
accountof the resultsof systematicreviewsandstudiesthathavetakenmeasuresto
reducethe effects of biasesandchance.As a citizen,too, I wantthese principlesto
be respected more generally-by policy makers,practitioners,and the public-
38 THE ANNALS OF THE AMERICAN ACADEMY
than they are currently. However, to return to my starting point, uncertainty and
humility among policy makers, practitioners, and researchers are the precondi-
tions for wider endorsement of the approaches I have outlined. Sadly, these quali-
ties are too often in short supply.
References
Achara,S., B. Adeyemi, E. Dosekun, S. Kelleher,M. Lansley,I. Male, N. Muhialdin,L. Reynolds, I. Roberts,
M. Smailbegovic,and N. van der Spek.2001. Evidence based roadsafety:The DrivingStandardsAgency's
schools programme.Lancet 358:230-32.
Antiplatelet Trialists'Collaboration. 1988. Secondary prevention of vascular disease by prolonged anti-
platelet treatment. BMJ296:320-31.
Antman, E. M., J. Lau, B. Kupelnick, F. Mosteller, and T. C. Chalmers. 1992. A comparison of results of
meta-analysesof randomizedcontroltrialsand recommendationsof clinicalexperts.Journalofthe Amer-
ican Medical Association 268:240-48.
Auerbach,A. W. 1978. The role of the therapeutic community "StreetPrison"in the rehabilitationof youth-
ful offenders. UniversityMicrofilmsno. 78-01086. Doctoral diss., George WashingtonUniversity,Wash-
ington, DC.
Boruch, R. 1997. Randomizedexperimentsfor planning and evaluation:A practical guide. Thousand Oaks,
CA: Sage.
Boruch, R., and F. Mosteller.2002. Overviewand new directions. In Evidence matters:Randomisedtrials in
education research,edited by F. Mosteller and R. Boruch, 1-14. Washington,DC: BrookingsInstitution.
Britton,A., M. McKee, N. Black, K. McPherson, C. Sanderson,and C. Bain. 1998. Choosing between ran-
domised and non-randomisedstudies:A systematicreview.Health TechnologyAssessment2 (13): 1-124.
Burns, T., M. Knapp,J. Catty,A. Healey, J. Henderson, H. Watt, and C. Wright.2001. Home treatment for
mental health problems: A systematic review. Health TechnologyAssessment 5 (15): 1-139.
Campbell, D. T. 1969. Reforms as experiments. American Psychologist 24:409-29.
Chalmers, I. 1993. The Cochrane Collaboration: Preparing, maintaining and disseminating systematic
reviewsof the effects of health care. In "Doing more good than harm:The evaluationof health care inter-
ventions,"edited by K. S. Warrenand F. Mosteller.Annalsof the New YorkAcademyof Sciences 703 (spe-
cial iss.): 156-63.
.1995. Whatdo I want fromhealth researchand researcherswhen I am a patient?BMJ310:1315-18.
2000. A patient'sattitude to the use of researchevidence to guide individualchoices and decisions in
health care. Clinical Risk 6:227-30.
. 2001. Invalid health informationis potentially lethal. BMJ 322:998.
.2002. Whywe need to knowwhetherprophylacticantibioticscan reduce measles-relatedmorbidity.
Pediatrics 109:312-15.
Chalmers, I., L. V. Hedges, and H. Cooper. 2002. A brief history of research synthesis. Evaluation and the
Health Professions25:12-37.
Chalmers, I., and R. Lindley. 2000. Double standardson informed consent to treatment. In Informedcon-
sent in medical research, edited by L. Doyal and J. S. Tobias,266-75. London: BMJ Publications.
Clarke, M., P. Alderson, and I. Chalmers.2002. Discussion sections in reports of controlled trialspublished
in general medical journals.Journal of the American MedicalAssociation 287:2799-801.
Clarke, R. V. 1997. Situational crime prevention: successful case studies. 2d ed. New York:Harrow and
Heston.
Clarke, R. V.G., and D. B. Cornish. 1972. The controlledtrial in institutionalresearch-Paradigm or pitfall
for penal evaluators? Home Office Research Study no. 15. London: Her Majesty'sStationeryOffice.
Cooper, H. M. 1982. Scientificprinciplesfor conducting integrativeresearchreviews.Reviewof Educational
Research 52:291-302.
Cornish, D. B., and R. V. G. Clarke. 1975. Residentialtreatmentand its effects on deliquency. Home Office
Research Study no. 32. London: Her Majesty'sStationeryOffice.
TRYING TO DO MORE GOOD THAN HARM 39
Dobash, R. E., and R. P. Dobash. 2000. Evaluating criminaljustice interventions for domestic violence.
Crime & Delinquency 46:252-70.
Freed, C. R., P.E. Greene, R. E. Breeze, W.Y.Tsai,W. DuMouchel, R. Kao, S. Dillon, H. Winfield, S. Culver,
J. Q. Trojanowski,D. Eidelberg, and S. Fahn. 2001. Transplantationof embryonicdopamine neurons for
severe Parkinson'sdisease. New EnglandJournal of Medicine 344:710-19.
Gilbert, R. 1994. The changing epidemiology of SIDS. Archives of Disease in Childhood 70:445-49.
Glass, G. V. 1967. Primary,secondary and meta-analysisof research. Educational Researcher 10:3-8.
Goldthorpe, J. H. 2001. Causation, statistics, and sociology.European SociologicalReview 17:1-20.
Graebsch, C. 2000. Legal issues of randomizedexperiments on sanctioning.Crime & Delinquency 46:271-
82.
Gueron, J. 2002. The politics of randomassignment:Implementing studies and affecting policy.In Evidence
matters:Randomisedtrials in educationresearch,edited by F. Mosteller and R. Boruch, 15-49. Washing-
ton, DC: BrookingsInstitution.
Hammersley, M. 2001. On "systematic"reviews of research literatures:A "narrative"response to Evans &
Benefield. British Educational ResearchJournal 27:543-54.
.2002. Systematicor unsystematic,is that the question? Some reflections on the science, art,and poli-
tics of reviewingresearchevidence. Textof a talk given to the Public Health Evidence Steering Groupof
the Health Development Agency, October, in London.
Juni, P., D. G. Altman, and M. Egger. 2001. Systematicreviews in health care: Assessing the qualityof con-
trolled clinical trials. BMJ 323:42-46.
Kippax,S. 2003. Sexualhealth interventionsare unsuitable for experimentalevaluation.In Effective sexual
health interventions:Issues in experimentalevaluation, edited by J. Stephenson, J. Imrie, and C. Bonell,
17-34. Oxford:Oxford University Press.
Kleijnen, J., P. Gotzsche, R. H. Kunz, A. D. Oxman, and I. Chalmers. 1997. So what's so special about
randomisation?In Non-randomreflectionson health services research:On the 25 anniversaryof Archie
Cochrane'sEffectiveness and efficiency, edited by A. Maynardand I. Chalmers, 93-106. London: BMJ
Books.
Konnerup, M. 2002. The three main pillars of the Campbell Collaboration.Presented at Nordic Campbell
Center InaugurationSeminar,12 November,in Copenhagen, Denmark. Retrieved 18 January2003 from
from http://www.nordic-campbell.dk/MereteKonnerupsforedrag4/index.htm.
Kunz, R., and A. D. Oxman. 1998. The unpredictabilityparadox:Review of empirical comparisonsof ran-
domised and non-randomisedclinical trials. BMJ 317:1185-90.
Kunz, R., G. Vist, and A. D. Oxman.2003. Randomisationto protect againstselection bias in healthcaretrials
(Cochrane methodology review). In The Cochrane Library, iss. 1. Oxford:Update Software.
Learmonth,A. M., and N. J. Watson. 1999. Construingevidence-based health promotion:perspectives from
the field. Critical Public Health 9:317-33.
Macintyre,S., and M. Petticrew.2000. Good intentions and received wisdom are not enough.Journalof Epi-
demiology and Community Health 54:802-3.
MacLehose, R. R., B. C. Reeves, I. M. Harvey,T. A. Sheldon, I. T. Russell, and A. M. Black.2000. A system-
atic review of comparisonsof effect sizes derived from randomisedand non-randomisedstudies. Health
TechnologyAssessment4 (34): 1-154.
McDonagh, M. S., P. F. Whiting, P. M. Wilson, A. J. Sutton, I. Chestnutt,J. Cooper, K. Misso, M. Bradley,E.
Treasure,and J. Kleijnen. 2000. Systematic review of water fluoridation.BMJ321:855-59.
Medical ethics Should medicine turn the other cheek? 1990. Lancet 336:846-47.
National Institute of Child Health and Human Development. 2000. Report of the National Reading Panel.
Teachingchildren to read: An evidence-basedassessmentof the scientific researchliteratureon reading
and its implicationsfor reading instruction. NIH Publication no. 00-4769. Washington, DC: Govern-
ment Printing Office.
NHS Centre for Reviews and Dissemination, Yorkand School of Sociology & Social Policy, Nottingham.
1999. Therapeutic community effectiveness. CRD Report 17. Retrieved 18 January 2003 from
www.york.ac.uk/inst/crd/.
Oakley,A. 1999. Paradigmwars. InternationalJournal of Social Research Methodology2:247-54.
.2000. Experimentsin knowing. Oxford:Polity Press.
40 THE ANNALS OF THE AMERICAN ACADEMY