The title promises a chapter about methods, so a confession is in order. Here, as everywhere, my concerns are substantive,
not methodological. Still, what one wants to learn and how one ought to go about learning it are intertwined. So, I propose
to bring out the logic of the survey experiment by presenting a classification of survey experiment designs. Specifically,
I distinguish three designs: manipulative, permissive, and facilitative. The distinctions among the designs turn on the hypotheses
being tested, not the operations performed, and, above all, on the role of predispositions. The first design aims to get people
to do what they are not predisposed to do; the second to allow them to do what they are predisposed to do, without encouraging
them; and the third to provide them with a relevant reason to do what they already are predisposed to do. Against the background
of this threefold classification, I want to comment briefly on some issues of causal inference and external validity and then
conclude by offering my own view on the reasons for the explosive growth in survey experiments in the study of public opinion.
Experiments have been part of public opinion surveys for many years, but they took the form of the so-called
split ballot. The questionnaire would be printed in two versions, and test questions would appear in each that were identical
in every respect but one. A more procrustean design is difficult to imagine. So, it is the more compelling testimony to the
ingenuity of researchers that they still managed to learn a good deal.
1 All the same, what they learned, although of advantage for the applied side of public opinion
research, was of less value for the academic side. Indeed, if one wanted to be waspish, then one could argue that this first
generation of survey experiments did more harm than good. By appearing to show that even trivial changes in question wording
could produce profound changes in responses, they contributed to a zeitgeist that presumed that citizens did not really have
genuine attitudes and beliefs. And, by the sheer repetitiveness of the split ballot design, they reinforced in the minds of
several generations of subsequent researchers that survey experiments had to fit the straight jacket of the two – and only
two –
conditions.
Partly because the split ballot was procrustean, and partly because my interest is substantive rather than methodological,
the idea that survey experiments could be a useful tool did not enter my head. The idea came to me by a more circuitous route.
The coffee machine at
the Survey Research Center (SRC) at the University of California, Berkeley, was located on the second floor. Because there
was only one machine, whoever wanted coffee had to go there. One day, in 1983, a cup of coffee was just what I wanted. The
line was long and directly behind me was Merrill Shanks. He was pumped up, so I asked him what was happening. Merrill is basketball
tall; I, less so, it would be fair to say. It must have been quite a sight, Merrill towering over me, gesticulating with excitement,
explaining that he had succeeded in writing a general-purpose,
computer-assisted interviewing program and illustrating with (both physical and mental) gusto the measure of the breakthrough.
2 I was thrilled for my friend's achievement, even if quite uninterested in the achievement itself. Computer-assisted interviewing
passed right through – and out of – my mind.
A year later, we took our children to spend a year in Toronto, living at my in-laws’ house, so that they would know their
grandparents, and their grandparents would know them – not as children parachuted in from California for a brief stay, with
their grandmother placing vats of candy by their bedsides, but as a family living together. It seemed like a good idea, and
once again I learned the danger of good ideas. Our children were heartbroken at returning to California. There was an upside,
however. Living with one's in-laws, however welcoming they are, is an out-of-equilibrium experience. I mention this only because
it says something about the social psychology of discovery. I do not believe that I would have had the breakthrough idea about
computer-assisted survey experiments were it not for the sharp and long break with everyday routine. Among other things, it
allowed the past to catch up with the present.
As a child, I went to a progressive summer camp. After a day of games on land and water, we would be treated to a late-afternoon
lecture in the rec hall on issues of social importance.
One of the lessons that we were taught was that discrimination and prejudice are quite different things. Prejudice is how
others feel about us (i.e., Jews), whereas discrimination is how others treat us. Although prejudice is a bad thing, how others
feel about us is not nearly as important as how they treat us. Covenants against Jews buying property in “protected” areas,
bans on membership in clubs, and quotas on university admission were the norm then.
3 But between then and now, the memory of the lecture on the difference between prejudice and discrimination would regularly
recur, and I would just as routinely be struck by the frustrating irony that I had enlisted in a vocation, survey research,
that could study prejudice (attitudes) but not discrimination (action). That persisting frustration, I believe, was behind
the idea that struck me on my walk with such force.
Here was the idea. The first computer-assisted interviewing program was purpose built for
question sequencing. Depending on the answer that a respondent gave to the first question in a series, the interviewer's screen
would automatically light up with the next appropriate question. In turn, depending on the answer that she gave to the second
question, the screen would light up with the next appropriate question, and so on. The depth of Shanks’ achievement, though,
was that he had transformed the opinion questionnaire into a computer program that could be put to many uses. His was question
sequencing, whereas mine was randomized
experiments.
I saw that, with one change –
inserting a random operator to read a computer clock – computer-assisted interviewing could provide a platform for randomized
experiments. Depending on the value of the random operator, questions could be programmed to appear on an interviewer's monitor,
varying the wording, formatting, and order. The procedure would be effortless for the interviewer. She need only ask the form
of the question that appeared on her monitor. And it would be invisible to the respondent. The fact that there were multiple
versions of the question, randomly administered, would be invisible to the interviewees because they would only be asked one.
4
Voila! There was the broad answer to the summer camp lecture on the distinction between prejudice and discrimination. Ask
a randomly selected set of respondents how much help the government should give to a white American who lost her job in finding
another. Ask the others exactly the same question, except assume that it is a black American who has been laid off. If more
white Americans back a claim to government assistance if the beneficiary is white, then we are capturing not only how they
feel about black Americans but also how they treat them.
That was the idea – and I remember the street that I was on and the house that I was looking at when I had it. And absolutely
nothing would have come of it but for Tom Piazza. Although Tom and I had seen each other around the halls of the SRC for years,
the main thing we knew about each other is that we shared an interest in the analysis of racial attitudes
(see Apostle et al.
1983). Blanche DuBois relied on the kindness of strangers. I have relied on their creativity and character. Tom was the one who
made computer-assisted randomized experiments work. Every study that I have done since, we have done together, regardless
of whether his name appeared on the project.
Childhood memories, disruption of routines, social science as a collaborative enterprise, technology as door opening, research
centers as institutionalized sources of ecological serendipity – those are the themes of the first part of my story on the
logic of discovery. The theme of the second part of my story is a variation on Robert Merton's (1973) classic characterization
of the communist – his word – character of science.
5
Tom and I had a monopoly position. Rather than take advantage of Merrill's breakthrough, the Institute for Social Research (ISR) at the University of Michigan and the National Opinion Research Center (NORC) at The University of Chicago attempted – for years – to write their own computer-assisted
program. It was a bad decision for them because they failed. An ideal outcome for us, you might think. Only studies done through
the Berkeley SRC could exploit the flexibility of computer-assisted interviewing in the design of randomized experiments,
which meant that we would have no competition in conducting survey experiments for years into the future.
Merton was right about the communist character of science, however. We would succeed, but we would do so alone. And if we
succeeded alone, we would fail. If other researchers could not play in our sandbox, then they would find another sandbox to
play in. Our work would always be at the margins.
The properly communist character of survey research showed up in a second way.
Public opinion surveys are expensive, so not many researchers got a chance to do them. Then Warren Miller effected the biggest-ever
structural change in the study of public opinion and elections. Consistent with Merton's doctrine of communism, Miller made
the data of the flagship voting studies nationally shared scientific property. In my judgment, there cannot be enough parades
in his honor. It was quantitative analysis that shot ahead, however, not the design of surveys.
6 The
American National Election Studies offered the opportunity for some innovation in measurement. But its overriding obligation
was to time series. Continuity of design was the primary value; innovation in design was secondary.
I had had my chance to come to bat in designing a study, actually two studies.
7 The first article that we succeeded in publishing using randomized experiments gave me the idea. The article was built on
the analysis of two experiments. Yet each experiment was only a question, admittedly a question that came in many forms, but
at the end of the day, only a question, which is to say that the experiment took only about thirty seconds to administer.
It then came to me that an interview of standard length could be used as a platform for multiple investigators. Each would
have time for two to four experiments; each would have access to a common pool of right hand–side variables; each would be
a principal investigator. If their experiments were a success, they would be a success. If not, they would have had a chance
to swing at the ball.
This idea of a shared platform for independent studies was the second-best design idea on my score card. It made it possible
for investigators, in the early stages of their careers, to do original survey research without having to raise the money.
8 But how to identify who should have the chance? A large part of the motivation is that very few had had an opportunity to
distinguish themselves through the design of original studies. My solution: I shamelessly solicited invitations to give talks
at any university that would have me in order to identify a pool of possible participants. I then invited them to write a
proposal, on the understanding that their idea was theirs alone, but that the responsibility for making the case to
the National Science Foundation (NSF) was mine. My sales pitch was, “We will do thirteen studies for the price of one.” That
was the birth of the
Multi-Investigator Project. It is Karen Garret, the director of the project, who deserves the credit for making the studies
a success.
I have one more personal note to add. The Multi-Investigator Project ran two waves. The day that I received the grant from
NSF for the second wave, I made a decision. I should give up the project. Gatekeepers should be changed, I had always believed,
and that applied to me, too. Diana Mutz and Arthur Lupia were the obvious choices. As the heads of Time-sharing Experiments in the Social Sciences (TESS), they transformed the Multi-Investigator Project. To get some order
of the magnitude of the difference between the two platforms, think of the Multi-Investigator Project as a stagecoach and
TESS as a Mercedes-Benz truck. Add the support of the NSF, particularly through the Political Science Program, the creativity
of researchers, and the radical lowering of costs to entry through cooperative election studies, and survey experiments have
become a standard tool in the study of public opinion and voting surveys. There is not a medal big enough to award Lupia and Mutz that would do justice to their achievements.
What is good fortune? Seeing an idea of yours travel the full arc, from being viewed at the outset as ridiculous to becoming
in the end commonplace,
9 and my sense of the idea
has itself traveled an arc. Originally, I saw it as a tool to do one job. Gradually, I came to view it as a tool to do
another.
Apostle, Richard A., Charles Y. Glock, Thomas Piazza, and Marijane Suelze. 1983. The Anatomy of Racial Attitudes. Berkeley: University of California Press.
Bartels, Larry, and Henry E. Brady. 1993. “The State of Quantitative Political Methodology.” In The State of the Discipline II, ed. Ada Finifter. Washington, DC: American Political Science Association, 121–62.
Chong, Dennis, and James N. Druckman. 2007a. “Framing Public Opinion in Competitive Democracies.” American Political Science Review 101: 637–55.
Chong, Dennis, and James N. Druckman. 2007b. “Framing Theory.” Annual Review of Political Science 10: 103–26.
Converse, Philip E. 1964. “The Nature of Belief Systems in Mass Publics.” In Ideology and Discontent, ed. David E. Apter. New York: Free Press, 206–61.
Dawson, Michael C. 2001. Black Visions. Chicago: The University of Chicago Press.
Druckman, James N. 2001a. “Using Credible Advice to Overcome Framing Effects.” Journal of Law, Economics, & Organization 17: 62–82.
Druckman, James N. 2001b. “On the Limits of Framing Effects.” Journal of Politics 63: 1041–66.
Druckman, James N. 2001c. “The Implications of Framing Effects for Citizen Competence.” Political Behavior 23: 225–56.
Druckman, James N. 2004. “Political Preference Formation.” American Political Science Review 98: 671–86.
Druckman, James N., Cari Lynn Hennessy, Kristi St. Charles, and Jonathan Weber. 2010. “Competing Rhetoric over Time: Frames versus Cues.” Journal of Politics 72: 136–48.
Freedman, David A., Roger Pisani, and Roger Purves. 1998. Statistics. New York: Norton, 3–11.
Gaines, Brian J., James H. Kuklinski, and Paul J. Quirk. 2007. “The Logic of the Survey Experiment Reexamined.” Political Analysis 15: 1–20.
Gibson, James L., and Amanda Gouws. 2003. Overcoming Intolerance in South Africa: Experiments in Democratic Persuasion. New York: Cambridge University Press.
Hacking, Ian. 1983. Representing and Intervening: Introductory Topics in the Philosophy of Natural Science. New York: Cambridge University Press.
Hurwitz, Jon, and Mark Peffley.1998. Perception and Prejudice: Race and Politics in the United States. New Haven, CT: Yale University Press.
Jackman, Simon, and Paul M. Sniderman. 2006. “The Limits of Deliberative Discussion: A Model of Everyday Political Arguments.” Journal of Politics 68: 272–83.
Kuklinski, James H., Paul M. Sniderman, Kathleen Knight, Thomas Piazza, Philiip E. Tetlock, Gordon R. Lawrence, and Barbara Mellers. 1997. “Racial Attitudes Toward Affirmative Action.” American Journal of Political Science 41: 402–19.
Merton, Robert K. 1973. The Sociology of Science. Chicago: The University of Chicago Press.
Nelson, Thomas E., and Donald R. Kinder. 1996. “Issue Frames and Group-Centrism in American Public Opinion.” Journal of Politics 58: 1055–78.
Riker, William H. 1996. The Strategy of Rhetoric. New Haven, CT: Yale University Press.
Schuman, Howard, and Stanley Presser. 1981. Questions and Answers in Attitude Surveys. New York: Academic Press.
Sniderman, Paul M., and Edward G. Carmines. 1997. Reaching beyond Race. Cambridge, MA: Harvard University Press.
Sniderman, Paul M.,
Joseph F. Fletcher,
Peter Russell, and
Philip E. Tetlock.
1996.
The Clash of Rights: Liberty, Equality, and Legitimacy in Pluralist Democracies.
New Haven, CT:
Yale University Press.
Sniderman, Paul M., and Thomas Piazza. 1993. The Scar of Race. Cambridge, MA: Harvard University Press.
Sniderman, Paul M., and Thomas Piazza. 2002. Black Pride and Black Prejudice. Princeton, NJ: Princeton University Press.
Sniderman, Paul M., Philip E. Tetlock, and Laurel Elms. 2001. “Public Opinion and Democratic Politics: The Problem of Non-attitudes and Social Construction of Political Judgment.” In Citizens and Politics: Perspectives from Political Psychology, ed. James H. Kuklinski. New York: Cambridge University Press, 254–84.
Sniderman, Paul M., and Sean M. Theriault. 2004. “The Structure of Political Argument and the Logic of Issue Framing.” In Studies in Public Opinion, eds. Willem Saris and Paul M. Sniderman. Princeton, NJ: Princeton University Press, 133–65.
Stoker, Laura. 1998. “Understanding Whites’ Resistance to Affirmative Action: The Role of Principled Commitments and Racial Prejudice.” In Perception and Prejudice: Race and Politics in the United States, eds. Jon Hurwitz and Mark Peffley. New Haven, CT: Yale University Press, 135–70.
Webb, Eugene, J., Donald T. Campbell, Richard D. Schwartz, and Lee Sechrest. 1996. Unobtrusive Measures: Nonreactive Research in the Social Sciences. New York: Rand McNally.
Zaller, John R. 1992. The Nature and Origins of Mass Opinion. New York: Cambridge University Press.