Sample Pages
Sample Pages
Contents
vii
2ND PAGES
Contents Copyright American Psychological Association
viii
2ND PAGES
Copyright American Psychological Association Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 859
ix
2ND PAGES
Copyright American Psychological Association
Contents
vii
2ND PAGES
Contents Copyright American Psychological Association
viii
2ND PAGES
Copyright American Psychological Association Contents
Part VI. Quantitative Research Designs Involving Single Participants or Units . . . . 745
Chapter 33. Single-Case Experimental Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 747
John M. Ferron, Megan Kirby, and Lodi Lipien
Chapter 34. Time Series Designs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 763
Bradley J. Bartos, Richard McCleary, and David McDowall
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 873
ix
2ND PAGES
Copyright American Psychological Association
Contents
vii
2ND PAGES
Contents Copyright American Psychological Association
viii
2ND PAGES
Copyright American Psychological Association Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 617
ix
2ND PAGES
Copyright American Psychological Association
Harris Cooper, PhD, is the Hugo L. Blomquist Professor, Emeritus, in the Department
of Psychology and Neuroscience at Duke University. His research interests follow
two paths. The first concerns research synthesis and research methodology. His book
Research Synthesis and Meta-Analysis: A Step-by-Step Approach (2017) is in its fifth edition.
He is the coeditor of the Handbook of Research Synthesis and Meta-Analysis (3rd ed.; 2019).
In 2007, Dr. Cooper was the recipient of the Frederick Mosteller Award for contributions
to research synthesis methodology given by the International Campbell Collaboration.
In 2008, he received the Ingram Olkin Award for distinguished lifetime contribution to
research synthesis from the Society for Research Synthesis Methodology. Dr. Cooper
also studies the application of social and developmental psychology to education policy.
In particular, he studies the relationship between time and learning.
Dr. Cooper chaired the first American Psychological Association (APA) committee
that developed guidelines for information about research that should be included in
manuscripts submitted to APA journals. In 2011, he published a book on the topic,
Reporting Research in Psychology: How to Meet Journal Article Reporting Standards. In 2020,
Dr. Copper published a revised second edition of the book as Reporting Quantitative
Research in Psychology: How to Meet APA Style Journal Article Reporting Standards.
Dr. Cooper currently serves as the editor of American Psychologist, the flagship journal
of APA. He served as editor for the Psychological Bulletin from 2003 through mid-2009.
Psychological Bulletin is in the top five social science journals in total citations and impact
factor. He was the chair of the APA Council of Editors in 2006 and was a member of the
committee that revised the APA Publication Manual (2010). In 2012, Dr. Cooper became
the inaugural coeditor of the Archives of Scientific Psychology, APA’s first open methods,
collaborative data sharing, open access journal. He remained as editor until 2015.
From 2009 to 2015, Dr. Cooper served as the chief editorial advisor for APA’s journal
publishing program. In this role, he served as a resource to the editors of APA’s 70+ journals
as well as the mediator of disputes between editors and authors and between authors
and authors. Dr. Cooper’s book Ethical Choices in Research: Managing Data, Writing
Reports, and Publishing Results in the Social Sciences (2016) draws from the experience.
The book goes beyond the proper treatment of human research subjects to examine
frequently neglected ethical issues that arise after data have been collected.
Dr. Cooper served as the chair of the Department of Psychology and Neuroscience
at Duke University from 2009 to 2014. He also served as chair of the Department of Psycho-
logical Sciences at the University of Missouri and director of Duke University’s Program
in Education. From 2017 to 2018 he served as the dean of social sciences at Duke.
xiii
2ND PAGES
Copyright American Psychological Association
xv
2ND PAGES
About the Associate Editors Copyright American Psychological Association
the same set of data is analyzed using each of five leading qualitative methodologies.
For the past couple of decades, her research has focused on discursive analyses of service
providers’ and service users’ accounts of depression and the use of antidepressants.
She is coeditor (with Janet M. Stoppard) of Situating Sadness: Women and Depression
in Social Context.
Dr. McMullen has been engaged with building the Society for Qualitative Inquiry
in Psychology (SQIP; a section of Division 5 [Quantitative and Qualitative Methods]
of APA) into a vibrant scholarly society since its earliest days. She was a member
of the Executive Committee from 2013 to 2021, served as its president in 2015–2016,
and was elected SQIP Section Representative to the Division in 2018 for a 3-year term.
In this latter capacity, she headed the multiyear process of developing an organizational
structure for SQIP, codifying it in a set of bylaws, and harmonizing the SQIP bylaws
with those of Division 5.
While working as a university professor, Dr. McMullen took on many leadership roles,
including director of clinical training for the graduate program in clinical psychology
at the University of Saskatchewan (1988–1994; 1995–1997), head of the Department
of Psychology (1997–2002; 2003–2006), university leader for the Social Sciences and
Humanities Research Council of Canada (2011–2013), and acting vice-dean for social
sciences (2012–2014). She served as the elected faculty member on the university’s
Board of Governors from 2004 to 2010.
Dr. McMullen’s contributions have been recognized by Division 5 of the APA
(2021 Distinguished Contributions to Qualitative Inquiry Award), the Canadian Psycho-
logical Association (2003, Fellow; 2012 Distinguished Member, Section for Women
and Psychology), and the Saskatchewan Psychological Association (1994 Award for
Outstanding and Longstanding Service to the Profession).
A. T. Panter, PhD, is the senior associate dean for undergraduate education and
a professor of psychology in the L. L. Thurstone Psychometric Laboratory at University
of North Carolina at Chapel Hill. She received her BA from Wellesley College in 1985
and her PhD from New York University in 1989. She is past president of the American
Psychological Association’s (APA’s) Division on Quantitative and Qualitative Psychology.
As a quantitative psychologist, she develops instruments, research designs, and data-
analytic strategies for applied research questions in higher education, personality, and
health. Dr. Panter serves as a program evaluator for UNC’s Chancellor’s Science Scholars
Program, a multisite adaptation of the successful Meyerhoff Program. She was also
principal investigator for The Finish Line Project, a $3 million First in the World grant
from the U.S. Department of Education that systematically investigated new supports
and academic initiatives, especially for first-generation college students. Her books
include the APA Dictionary of Statistics and Research Methods (2013), The APA Handbook
of Research Methods in Psychology (1st ed.; 2012), the Handbook of Ethics in Quantitative
Methodology (2011), The SAGE Handbook of Methods in Social Psychology (2004),
and volumes on program evaluation for HIV/AIDS multisite projects.
At the undergraduate level, she teaches statistics, research methods, and a first-year
seminar on communicating research results to others (“Talking About Numbers”).
At the doctoral level, she teaches courses in research design, classical and modern
approaches to instrument/survey design, and test theory and multivariate methods.
Dr. Panter has received numerous awards for her teaching and mentoring, including the
xvi
2ND PAGES
Copyright American Psychological Association About the Associate Editors
Tanner Award, the J. Carlyle Sitterson Award, a Bowman and Gordon Gray Distinguished
Professorship, and APA’s Jacob Cohen Award for Distinguished Teaching and Mentoring.
She is an APA Fellow, a member of the Graduate Record Examination Advisory Board,
a former member of a Social Security Administration advisory panel related to disability
determination, and a member of APA’s Committee on Psychological Testing and
Assessment. She regularly provides services for federal agencies, national advisory
panels, and editorial boards.
Dr. Panter has been a member of the university faculty since 1989. As senior associate
dean, she oversees all of the College’s programs for undergraduate education: academic
advising, undergraduate research, student success and academic counseling (including
Learning Center, the Writing Center, Peer Mentoring, Carolina Firsts, Summer Bridge),
undergraduate curricula (including the general education curriculum), instructional
innovation, research and evaluation, Robertson Scholars Leadership Program, and Honors
Carolina (including distinguished scholarships). Among her active work on campus,
she helped design and implement the IDEAS in Action general education undergraduate
curriculum, developed student learning outcomes for the university system, implemented
a new holistic Thrive academic advising approach, address legal mandates related to the
use of race/ethnicity in undergraduate admissions decisions, developed visualizations
for key university data, and increased the number and type of high impact academic
experiences for all undergraduate students.
David Rindskopf, PhD, is a Distinguished Professor at the City University of New York
Graduate Center, specializing in research methodology and statistics. His main interests
are in Bayesian statistics, causal inference, categorical data analysis, meta-analysis, and
latent variable models. He is a Fellow of the American Statistical Association and the
American Educational Research Association and is past president of the Society of
Multivariate Experimental Psychology and the New York Chapter of the American
Statistical Association.
xvii
2ND PAGES
About the Associate Editors Copyright American Psychological Association
(where he received the Chancellor’s Award for Research and Creativity, the President’s
Award for Research and Creativity, and the SEC Faculty Achievement Award/Professor
of the Year), the National Institutes of Health (a MERIT Award, a Senior Scientist and
Mentoring Award, and a Mark Keller Lectureship), and Indiana University (Richard C.
Atkinson Lifetime Achievement Award). Throughout his career he has been heavily
involved in service to professional societies (e.g., he served as president of the Research
Society on Alcoholism and served on the APA’s Council of Representatives, Board of
Scientific Affairs, and Policy and Planning Board, chairing the latter two of these bodies).
Dr. Sher also has a long history of service to scholarly publications, serving as an
associate/field editor for both disciplinary journals (Clinical Psychological Science,
Journal of Abnormal Psychology, and Psychological Bulletin) and specialty journals
( Journal of Studies on Alcohol and Drugs and Alcoholism: Clinical and Experimental Research).
He recently served as acting editor for Clinical Psychological Science. His current work
is focused on attempting to define the “core” of addiction and developing methods to
improve the diagnosis of alcohol use disorders and other psychological disorders.
xviii
2ND PAGES
Copyright American Psychological Association
Introduction: Objectives of
Psychological Research and Their
Relations to Research Methods
METHODS OF KNOWING
The American philosopher Charles Peirce (1839–1914) claimed that we use five different
ways to decide what we believe is true about our world (Feibleman, 1969). First, we believe
some things are true because authorities we trust tell us so. Sometimes, we know these
authorities personally, such as our parents and teachers. Sometimes, they are very distant
from us, such as the writers of ancient religious texts. Other times, authorities are less
far removed but still not personally known, for example, the authors in a handbook
on research methods.
Second, we know things are true because we have heard them repeated many times.
Peirce called this the method of tenacity, or the a priori method. Here, something is
believed because it has always been believed (longevity) or because we have heard
it repeated many times. We could include in this method the commonsense adages
with which we are all familiar, such as “birds of a feather flock together” or “a stitch
in time saves nine.”
Third, we observe or experience things ourselves and our senses tell us they are true.
The sun warms things up, for example.
Fourth, we know that some things are true because they can be logically derived;
they are the product of rational analysis. Without getting into the formalities of logical
My sincere thanks go to the five associate editors of this handbook, Marc N. Coutanche, Linda M. McMullen, A. T. Panter,
David Rindskopf, and Kenneth J. Sher. The latter three also served as editors on the first edition. They deserve equal
billing with me as editors. They also provided feedback on this introduction, although any mistakes remain my doing.
A special note of thanks goes to Kristen Knight, APA project editor, for her diligence and organizational efforts.
xxv
2ND PAGES
Introduction Copyright American Psychological Association
deduction, if a premise known to be true tells us that “all males have an Adam’s apple,”
and we observe a men’s intercollegiate fencing match, then logic dictates we believe that
these fencers have Adam’s apples under their masks.
The problem with each of these first four methods of knowing is that they are
fallible. Two trusted authorities can disagree, suggesting that one (at least) must be wrong.
Tenacious beliefs can lead us astray because conditions change over time, or what seems
like common sense is not so sensible after all (remember, in addition to birds of a feather
flocking together “opposites attract,” and although a stitch in time may be frugal,
“haste makes waste”). Our senses can deceive us, for example, through the application
of different frames of reference, as demonstrated by optical illusions. Are the soccer
balls in Figure 1 the same size?
Finally, a logical deduction is based on the validity of the premises, which may be
wrong. Or, the logic itself may be faulty even though the premises are true, as when
we affirm the consequent (“All male fencers have an Adam’s apple,” and “Spencer has
an Adam’s apple”; therefore, “Spencer is a fencer”).
Peirce’s final method of knowing was the scientific method. We can think of the
scientific method as a combination of observation and rational analysis, or observation
using a set of logical rules that should lead to veridical conclusions about the world.
Peirce expected the scientific method to lead to knowledge that was exactly the same for
every person, uninfluenced by idiosyncratic frames of reference. He held out the hope
for truly objective knowledge.
xxvi
2ND PAGES
Copyright American Psychological Association Introduction
1 To be more precise, in the tradition of null hypothesis testing the scientist might say, “I propose an alternative to the
hypothesis that the two balls have equal diameter. I propose that the upper ball is larger. Now, let’s try to reject the equal
diameter hypothesis.”
xxvii
2ND PAGES
Introduction Copyright American Psychological Association
A self-critical posture requires that truth claims never be believed with absolute
certainty, only with greater or lesser certainty. This is what is most unique and, I think,
exciting about the scientific posture. It is also one tenet that binds the chapters in this
handbook. All of the chapter authors would agree that psychological scientists must
take a critical stance toward what they call “knowledge.”
xxviii
2ND PAGES
Copyright American Psychological Association Introduction
far apart on others. So, deciding which dimensions to prioritize has nontrivial implications
for where a method appears. In addition, prioritizing some dimensions over others can
border on arbitrary and be based on oversimplified characterizations of any methodology’s
capacities for guiding discovery. Many methods can be used for more than one purpose.
We have tried to put these “Swiss Army knives” of methodology in the toolbox com-
partment of their most frequent use. In addition, traditions of use within subdisciplines
dictated that some methods appear close together, even if grouping them defied the
logic of our dimensional analysis. And, some methods are so unique that they were hard
to place anywhere. These methods are no less important because of their singularity;
indeed, if they are the only way to answer a question, their uniqueness can make them
especially valuable.
So, as you scan the table of contents and contemplate our choices for clustering and
ordering the presentation of methods, I am certain that you will be perplexed by some
of our choices. Other schemes could fit equally well, or better. Below, I try to capture the
high-order dimensions that informed our placement of chapters, beginning with those
that relate to the earliest decisions that a researcher makes when choosing methods.
Clearly, this searcher’s method does not fit the objective. Similarly, psychological
researchers must choose methods that fit the research question that they want to answer,
not the method that is available or that they know best. No matter how luminous a
method is, if the questions it can answer do not correspond to the knowledge sought,
the researcher will remain in the dark.
You could think of this handbook as a collection of streetlights. Each method
contained herein is meant to help you shed light on thought, feeling, and behavior over
a different expanse and from a different angle. As I alluded to, another frequent metaphor
compares research methods with a toolbox. Here, methodology provides the hammers,
screwdrivers, wrenches, and rulers that psychological researchers use when they ply
their trade.
You will read repeatedly in the chapters that follow that your first task as a psycho-
logical researcher is to pick the method best suited to answer the question that
motivates you. You will be told not to search where the light is or bang a nail with
a screwdriver. Instead, you will learn to choose the method that best answers your
question. The contributors hope that this handbook will expand the topics that you
can illuminate and increase the size of your toolbox. We hope to provide you with new
ways to answer old questions as well as to raise new questions, perhaps ones you did
not realize could be asked.
xxix
2ND PAGES
Introduction Copyright American Psychological Association
At the broadest level, when choosing a method you make decisions about (a) what
measurement and data collection techniques best capture the thoughts, feelings, and
behaviors that interest you; (b) what research design best fits the question that you want
to answer; and (c) what strategies for data analysis best match the characteristics of your
measurements and design.
The simplest choice for organizing the presentation of material is the temporal sequence
in which you will make these decisions. This is roughly what we have done. So, the earliest
chapters in Volume 1, Part I, address the broadest questions related to research designs.
These involve both (a) which research designs are most appropriate for which question
and (b) how to think about the ethicality the research that address your question, and
(c) how to conduct research with participants drawn from more diverse populations.
Next, the chapters in Volume 1, Part II, help you with the research planning process,
including how to develop a testable hypothesis, find the pertinent literature, secure the
resources you need, and choose measures and what people to study. Part III of the first
volume describes the plethora of measurement techniques that psychologists most often
use to collect data and how to determine whether the measurement techniques that you
might choose are the best ones for your purpose. For this revised edition of the handbook,
significant changes were made to the discussions of chronometric and psychophysical
measures (Section 4) and the measures used in psychophysiology and neuroscience
(Sections 5 and 6).
Part IV contains chapter looking at different ways to assess the trustworthiness of
measures. These help you determine whether your measures can (or did) allow you find
the answers you sought.
In Volume 2, Parts I through VI, the chapters return to issues of research design.
They present for your consideration a panoply of options, further divided along more
nuanced distinctions in their objectives (discussed in the following section Interpretive
Inquiry, Description, and Causal Explanation).
Chapters on techniques for data analysis follow in Volume 3, Part I, again with special
attention to the fit between design, measurement, and analysis. Finally, issues and choices
you must consider when you write up your research to share with the community of
psychologists are discussed in the handbook’s concluding chapters, in Volume 3, Part II.
Interpretive Research
To choose between interpretive or descriptive research, you must also answer the
following question: Do you want to uncover the impetus to thoughts and actions that
exist for the actors themselves or do you have your own theory or perspective to guide
xxx
2ND PAGES
Copyright American Psychological Association Introduction
your data collection? Carla Willig (Chapter 1 in this volume) helps you decide whether
your question naturally fits in the former category, suggesting an interpretive inquiry
design. She suggests that qualitative research designs are most appropriate when the
interpretation of an event by participants is your goal or when your question falls into
one of these categories:
■ What does something feel like?
■ How is something experienced?
■ How do people talk about something and with what consequences?
■ How do people make sense of an experience?
■ How do they construct its meaning? What does this allow them to do or not to do?
To feel or not to feel?
■ How does a particular event unfold? How do participants experience the event?
What may be its consequences? For them and for others?
Willig makes the point that these approaches to psychological research are most
appropriate when the researchers do not want to impose their own (or someone else’s)
theory or perspective on the thoughts, feelings, or actions of the people that they are
studying. Rather, the researchers want to uncover the impetus to behavior that exists for
the actors themselves. Cultural anthropologists refer to this as using an emic approach
to describing behaviors and (conscious or unconscious) beliefs.
Qualitative designs (detailed in Volume 2, Part I) use a range of data, including
spoken or written narratives from interviews and informal conversations; archival data
contained in public records and private diaries; and visual data from photographs, film,
and video. Although these data are often obtained from relatively few participants, large-
scale, primarily quantitative, studies have increasingly employed a range of qualitative
methods to explore a diverse range of questions. In recent years, the desirability of using
multiple methods in research has become increasingly evident and is being adopted more
often. Volume 2, Part II contains new chapters that address the use of multiple methods
for collecting both qualitative and quantitative evidence in the same study.2
It is also possible to take an etic approach to research, or to provide a description of
an event. Here, the researchers’ theories and beliefs are applied to the situations that they
study. These forms of descriptive research often focus on a few specific characteristics
of events or individuals chosen a priori by the researcher. Participants are then sometimes
broadly sampled and they respond to questions developed by the researchers to investigate
particular aspects of the person or situation. Similar to qualitative research, the researchers
make no attempt to manipulate the participants’ circumstance. Similar to quantitative
research, the data collected will be in numerical form and examined through statistical
procedures.
2 For an interesting take on the similarities and differences between quantitative and qualitative approaches, see Shweder
(1996): “The true difference between the approaches is not over whether to count and measure but rather over what to
count and measure, and over what one actually discovers by doing so” (p. 179).
xxxi
2ND PAGES
Introduction Copyright American Psychological Association
3 See Cooper (2007) for my first presentation of this material on Hume (1739–1740/1978).
4 Note how this harkens back to the criticisms of Peirce.
xxxii
2ND PAGES
Copyright American Psychological Association Introduction
xxxiii
2ND PAGES
Introduction Copyright American Psychological Association
5But we can never be completely certain that the characteristic of the manipulation that we claim is causal was the
productive element because our experimental and comparison conditions can be viewed as differing on many
characteristics. Hume’s ghost again.
6You would not want to do this in a real experiment because it confounds the particular stimuli with the order. In this
imaginary scenario, you must assume that order has no influence, so that differences between people are due to the
message.
xxxiv
2ND PAGES
Copyright American Psychological Association Introduction
16
14
12
Person 1
10
Attitude
Person 2
8
Person 3
6
Person 4
4 Person 5
2 Person 6
0
1 2 3 4 5 6 7 8 9 10
Messages
12
10
8
Attitude
0
1 2 3 4 5 6 7 8 9 10
Messages
FIGURE 2. Hypothetical graphs of six individuals’ change in attitude
(top panel) and group-averaged attitudes of those six people (bottom panel)
toward energy conservation after viewing 10 proconservation messages with
different themes.
The bottom panel of Figure 2 provides a graph of how the same data look if averaged
across the six people. A much different picture emerges. Looking at this graph, if you
were to assume the group average effect accurately reflected what was happening to each
individual person, you would say that after the second exposure, each person gradually
changed their perspective on conservation after viewing each message. After eight
exposures, no additional change took place. Clearly, this would be an incorrect char-
acterization of the process occurring at the level of the individual person. That said,
the group-averaged data could be used to describe how change occurred for the group
as a whole, as a single unit. Thus, the correct interpretation of the group-averaged data
would be to say that when a group was exposed to proconservation messages, the group
average attitude changed gradually after the second exposure, but there was no additional
change after eight exposures. This would be an accurate description of how the group
behaved over exposures, but it would not adequately describe any single member within
the group.
Whether you are interested in the individual or group-averaged effect depends on
the context in which your question is being asked. Sometimes the group average does
xxxv
2ND PAGES
Introduction Copyright American Psychological Association
represent the behavior of a single unit. So the bottom panel of Figure 2 is a description
of, say, how one city’s perspective might be described if the 10 messages were weekly
programs aired on a local television channel. Then, the city becomes the unit, not the
individuals who populate it. If your problem focuses on understanding how individual
units change over time, then the top panel of Figure 2 provides you with six replications
of the phenomena of interest. The bottom panel is irrelevant to your question. If your
focus is on how a group average changes over time, the bottom panel provides you with
one instance of this and the top panel is irrelevant.
Are you seeking an interpretation, description, or causal explanation for an event or relationship?
Do you want to uncover Do you have your own Do you want to test Do you want to manipulate
the impetus to thoughts theory or perspective to the implications of a a presumed cause to
and actions that exist for guide your data causal model? estimate its possible
the actors themselves? collection? effect?
Are you interested in understanding (a) how a single unit behaves or changes over
time or (b) what makes one group of people different from another group, on average?
FIGURE 3. Relations between research questions, research designs, and the organization
of parts in the APA Handbook of Research Methods in Psychology.
xxxvi
2ND PAGES
Copyright American Psychological Association Introduction
approach are detailed in Part I. Most of these designs also focus on questions that
involve describing the current state or change in an individual unit of interest or a small
sample of individuals. Part II looks at issues at the transition point between qualitative
and quantitative research, as well as issues that arise in research with particular popula-
tions and settings that go beyond their simple participation in research.
Volume 2, Parts III through VII, introduce designs that emphasize an etic (or theory-
specified), more quantitative approach to research. Volume 2, Part III, presents issues in
sampling for quantitative studies. Although the techniques described herein will be of
interest to all researchers, they would be of special interest to those who are conducting
descriptive research. So, for example, if you are collecting data on the impact of growing
up in a single-parent home, this section will assist you in planning your strategy for
sampling respondents, help you consider issues you will encounter in collecting data
from participants over time, and help you determine how you might use the internet.
Volume 2, Part IV, focuses on designs that build and test the implications of causal
models. You will find here approaches that differ in the type of data that they employ
(sometimes even simulated data) and in the assumptions that are made as part of the
data analysis (e.g., Bayesian modeling). Volume 2, Part V, focuses on research with
experimental manipulations, in which participants are deliberately treated differently.
Section 1 of Part V distinguishes these designs depending on how participants were
assigned to their experimental conditions. Section 2 of Part V describes many of the
unique problems that psychologists face when they conduct experimental research in
applied settings.
In Volume 2, Part VI, designs are introduced that focus on theory-testing questions,
rely heavily on quantification, and are used to study change in individual units. These
designs all require multiple measurements of the same dependent variable(s) over a
period of time. They can be used to study change that is either (a) naturally occurring,
for example, as when a researcher wants to describe how a person’s cognitive abilities
change as they age, or (b) purposively manipulated, as when a researcher examines
an older adult’s cognitive ability before and after an intervention that is meant to
improve memory.
The designs in Volume 2, Part VII, are labeled “Neuropsychology” and “Genetic
Methods in Psychology.” Here you will find designs for theory-driven research that derive
largely from the more biological end of psychology’s family tree.
Units of Analysis
As a science matures, it adopts, adapts, and invents new techniques for looking at the
world. Certainly, the ruler that we used to examine the optical illusion in Figure 1 was
not invented for that purpose alone. Rather, we realized its relevance to our problem
and commandeered it for our cause. And, as you look at the contents of Volume 1,
xxxvii
2ND PAGES
Introduction Copyright American Psychological Association
Parts III and IV, it will be obvious to you that this handbook describes an enormous
array of rulers. Some of these rulers were invented by psychologists, but many were
invented for other purposes to study phenomena of interest in other disciplines.
It is possible to think of the sciences as falling along a continuum that distinguishes
them according to the size of the things that they study or their unit of analysis or
investigation. So chemists, generally speaking, study things that are physically smaller
than the things studied by biologists, whose units of study often are smaller than those
studied by psychologists, whose units often are smaller than those studied by sociologists.
Of course, the overlap in topics of interest is great, so at the margins the distinction
between disciplines breaks down; it becomes difficult, if not impossible, to identify
where one scientific discipline ends and the next begins. A psychologist who studies
group identity is more likely to share an intellectual heritage with many sociologists
than with a psychologist who studies the role of neurotransmitters in depression,
whose work may be more akin to that of a neurobiologist.
Along with a blurring at the margins of disciplines comes the transfer of measurements
and methods between disciplines. Not surprisingly, then, in this handbook, you will
find measurement techniques (as well as research designs and statistical techniques)
with histories that locate their roots in numerous fields of study, including economics,
political science, sociology, anthropology, neurobiology, and genetics. This is a good
thing for our discipline. Psychologists have come to recognize that a complete picture
of any phenomenon requires that it be examined through multiple methods, applying
multiple rulers. To fully understand schizophrenia, for example, psychological scientists
might need to survey its prevalence in a population, examine family dynamics, observe,
interview, and test individuals with the disorder, conduct brain scans, and map genes.
Because of psychology’s interdisciplinary range, the array of methods covered in this
handbook is daunting. But the variety of methods that psychologists use is indicative of
our discipline’s strength and vitality. The authors of the handbook chapters are motivated
by a search for answers, no parochialism here. They share the belief that their own
method of choice cannot develop a complete picture of the world or, really, any discrete
phenomenon in it. Rather, each method supplies a small piece of the puzzle. It is only
when the puzzle pieces are fit together that a complete picture emerges.
Volume 1, Part III, of the handbook offers many different techniques of measurement.
The sections are roughly organized according to the size of their unit of analysis. It begins
with the largest units and proceeds to the smallest. So, Section 1 presents techniques
that measure peoples’ overt individual behavior, which are typically available for others
to view. Sections 2 and 3 largely describe measures for which people provide verbal or
written data about what they are thinking, what they are feeling, or how they behave.
Sections 4, 5, and 6 reduce the unit of analysis even further, to psychophysical and
psychophysiological measures and then to measures that are biological in nature.
The chapters in Volume 1, Part IV, help you answer a second question about your
measures: “How well does your chosen measure represent the variable that interests you?”
This question again requires you to consider fit, but now between a concept, or latent
variable, and the means that are used to measure it. Put simply, the variables involved in
psychological research need to be defined in two ways, conceptually and operationally.
Conceptual definitions describe qualities of the variable that are independent of time
and space and can be used to distinguish events that are and are not instances of the
concept. For example, a conceptual definition of aggression might be “behavior intended
xxxviii
2ND PAGES
Copyright American Psychological Association Introduction
to cause pain or harm.” Conceptual definitions can differ in breadth, that is, in the
number of events that they capture. So, if the terms pain and harm are interpreted
broadly, then aggression could include verbal as well as physical acts.
To relate concepts to concrete events, a variable must also be operationally defined.
An operational definition is a description of the observable characteristics that allows us
to determine whether a behavior or event represents an occurrence of the conceptual
variable. So, an operational definition of aggression might include “shouting, or vocal-
izations above a specified decibel level” if verbal aggression is included but not so if only
physical harm is included. The chapters in Volume 1, Part IV, present the criteria and
many of the techniques that psychological researchers use to assess whether a measure
is a good fit for a construct.
7 Of course, multiple measures of state anxiety are as desirable a feature of your study as multiple measures of aggression.
xxxix
2ND PAGES
Introduction Copyright American Psychological Association
rule out the irrelevant influences (misinterpretation by observers, social desirability, etc.)
on your three aggression measures as the cause of the relation. If results are inconsistent
across operations, having the three measures allows you to speculate on what the
important differences between operations might be and refine your theory about how
the two constructs are related.
There is another benefit to including multiple operations in research. Predictions are
often made from theories on the basis of presumed causal processes that include multiple
steps, or causal linkages. These are the focus of causal modeling studies (Volume 2, Part IV),
and they also pertain to other research designs. So, we might speculate that hearing
an alarm increases state anxiety by increasing uncertainty and physiological arousal.
In turn, uncertainty and arousal increase the likelihood of aggression. By including
a measure of arousal in your study, along with observed and self-reported aggression
measures, your study also tests this mediating mechanism.
Of necessity, the chapters in Volume 1, Part III, present their array of measurement
methods largely as discrete choices. And, in Volume 1, Part IV, the methods for appraising
a measure’s fit with the construct of interest largely address the adequacy of each measure
separately. But as you design a study, you should not think that you must choose one
measure or another. Instead, when you consider which measure best captures your
construct of interest, remember that no measure is perfect. And, to the extent reasonable,
the more measures—that do not share the same imperfections and that test more than
one linkage if a causal chain is hypothesized—the better.
xl
2ND PAGES
Copyright American Psychological Association Introduction
uncover the abstract, or latent, variables that underlie a set of observed, or manifest,
variables. For example, we might use an exploratory factor analysis to determine the
number of factors underlying a multi-item measure of aggression or a confirmatory
factor analysis (on the same instrument) to test the theory that the tendency toward
physical and verbal aggression are independent (i.e., knowing how likely people are to
hit you tells you nothing about how likely they are to call you a nasty name). Within
Volume 3, Part I, Sections 3 to 5, you will also find a distinction in analytic choices
depending on whether they pertain to (a) variables that are measured continuously or
(b) variables that place people (or other units) into classes. Some of these techniques
take into account the categorical nature of variables in your analyses, whereas others
help you to discover what these categorical distinctions among respondents might be.
So, to find the chapters of greatest interest to you in these sections, first you will
need to answer three questions: (a) Does your research problem distinguish independent
or predictor variables from dependent or outcome variables? (b) Are your variables
measured over time? (c) Are your variables continuous or categorical, either in how
they are conceptualized or how they are measured? Of course, in many instances your
answers to these questions will be complex. Within the same study, for example, you
might want to reduce the number of dependent variables by constructing a composite
(e.g., factor analyze the observation, self-report, and physiological measures of aggression)
and then use the composite in a regression analysis (one dependent variable) with other,
multiple indicators. Or, you will have some continuous and some class variables. As you
read the chapters, you will find that one great advance in statistical science has been the
recent development of sophisticated techniques that permit the integrated analysis of
data that vary in their characteristics and the number of variables involved.
Volume 3, Part 1, Section 6, presents methods for studies that take into account
(and capitalize on) the interdependence of responses from multiple participants who
are in interaction with one another. The units of analysis can be anywhere from a dyad
(e.g., a husband and wife) to a large group (e.g., a sports team, workers in an office).
Because the responses of individuals in these networks are dependent on the responses
of others, the data analysis strategy must take this into account. Equally important,
sometimes the research question focuses on the nature of the interdependence.
Finally, in Volume 3, Part I, Section 7, you will find three chapters that present some
of the special issues that arise, and some special statistical techniques that are used,
when researchers reanalyze or integrate data that was collected by others. In one instance,
secondary data analysis, you work with raw data that might have been collected for
another purpose (e.g., arrest rates and climate data in cities are used to test the relation
between crime and temperature). In the other instance, meta-analysis, the statistical
results of previous research becomes the raw data in a quantitative research synthesis.
xli
2ND PAGES
Introduction Copyright American Psychological Association
of their data not only, “Are these variables related, yes or no?” but also “How strong of
a relationship is there?” (Wilkinson & Task Force on Statistical Inference, 1999).
Prominent among the methods used to describe data is the estimation and inter-
pretation of effect sizes, or “the degree to which the phenomenon is present in the
population” (Cohen, 1988, p. 9). With the coming of age of effect size estimation came
the importance of understanding the difference between statistical significance and
clinical or practical significance. The latter requires extrastatistical interpretation of the
data. To assess practical significance, researchers (and others) must wrestle with the
question of how strong a relation needs to be before it can be deemed meaningful or
important (see Cooper, 2008).
The answer to this question always depends on the context in which the research takes
place. Cohen (1988) suggested some general definitions for small, medium, and large
effect sizes in the social sciences. In defining these adjectives, he compared different
average effect sizes that he had encountered across disciplines in the behavioral sciences.
However, Cohen did not intend his labels to serve as guides for the substantive inter-
pretation of relations by social scientists. Rather, he intended his rules to assist with power
analyses in planning future studies, a very different objective. Using Cohen’s definitions
to interpret the substantive importance of an effect size misapplies his work.
In fact, there is no fixed scale for the substantive interpretation of the size of a relation,
and there is no substitute for knowing the research context of the specific question.
Here is a simple example. Assume that we have the results of a study that evaluated an
intervention that was conducted with 200 participants, 100 each in the intervention
and control condition, and a dichotomous measure of success or failure. Using Cohen’s
definitions, an increase in success rate from 45% in the control condition to 55% in
the intervention condition would be considered a small effect (equivalent to r = .10
explaining 1% of the variance). However, what if this effect were found on a measure
of “suicides among previous attempters” and the intervention was access to an online
psychological services? Personally, I would not be inclined to label this effect “small,”
practically speaking. However, if the study measured whether previous suicide attempters
did or did not endorse the statement “life is worth living” after a year of daily psycho-
therapy “small effect” certainly would come to my mind.
In this example, I tried to demonstrate that the interpretation of effect sizes rests
heavily on (a) the intrinsic value placed on the outcome variable (how valuable is even
a small difference?) and (b) the cost of the intervention. Also, when interpreting the
magnitude of effects, it is informative to use contrasting elements that are closely related
to the topic at hand. For example, what other interventions have been used to prevent
suicide among previous attempters? If the hotline and several other interventions have
been tried and found to have no effect, suddenly the daily therapy effect starts to look
larger, worth pursuing further.
Effect sizes also need to be interpreted in relation to the methodology used in the primary
research. So, studies with more intensive treatments (e.g., more frequent therapy sessions),
more sensitive research designs (within-subject rather than between-subject), and measures
with less random error can be expected to reveal larger effect sizes, all else being equal.
Although null hypothesis testing is not ignored, the contents of this handbook clearly
demonstrate the shifting of emphasis from “yes or no?” to “how much?” questions.
In all of the chapters on data analysis, you will find a primary emphasis on estimating
and interpreting the magnitude of relations.
xlii
2ND PAGES
Copyright American Psychological Association Introduction
xliii
2ND PAGES
Introduction Copyright American Psychological Association
The final part of the handbook addresses issues related to open science. It includes
three chapters that address data management and the integrity of the research reporting
process. The chapters in Volume 3, Part II, look at how to plan for managing your data,
including documentation and storage so that it is easily understood by others (Chapter 25),
what practices in data analysis that should be avoided so that your results are more easily
replicable (Chapter 26), and a broader overview of ethical issues that arise in research
not associated with the treatment of research participants (Chapter 27). All these
chapters are new to the second edition of the handbook.
CONCLUSION
“Introductions” to edited works are required to address what the chapters that follow
have in common as well as how the differences between chapters are to be understood.
Writing such a chapter for a handbook as broad in scope as this one has required that
I touch on similarities at lofty levels of abstraction, such as the methods by which
people know things and the relation of science to democracy.
But I have been able as well to uncover some very down-to-earth examples of simi-
larities in the chapters. For example, as I have noted several times, a principle shared
by all of the authors is that the research methods you choose should be appropriate
to answer the question that you pose. This dictum seems almost too obvious to state.
Let us not fool ourselves, however. The opposing desire to use the tool you know even
if it’s not a perfect fit is often hard to resist. Hopefully, this handbook will expand your
toolbox so that this latter approach loses its appeal.
Describing the differences between chapters and how they can be understood has
presented an equally formidable challenge. It was easy to begin with the sequence
of method choices—assessing the ethics and feasibility of different approaches, then
choosing measures, a research design, statistical techniques, and ending with research
reports—although we know that in practice these choices are never as linear as they
appear in books.
Bringing an equally linear order to the array of research designs, measurements,
and analytic techniques available to psychological scientists was the most difficult task.
Different approaches to psychological research begin with different epistemic assump-
tions and then travel through subdisciplines with different traditions. Like the species
that we study, the methods used by psychological scientists defy simple categorization.
But this is a good thing (even if it causes trouble for editors). After all, if science is
humankind’s greatest achievement (and I think it is), then isn’t turning the lens of
science on ourselves the ultimate expression of our uniqueness?
Harris Cooper
Editor-in-Chief
References
American Psychological Association. (2020). Publication manual of the American Psychological
Association (7th ed.).
APA Presidential Task Force on Evidence-Based Practice. (2006). Evidence-based practice in psychology.
American Psychologist, 61(4), 271–285. https://doi.org/10.1037/0003-066X.61.4.271
Appelbaum, M., Cooper, H., Kline, R. B., Mayo-Wilson, E., Nezu, A. M., & Rao, S. M. (2018). Journal
article reporting standards for quantitative research in psychology: The APA Publications and
xliv
2ND PAGES
Copyright American Psychological Association Introduction
Communications Board task force report. American Psychologist, 73(1), 3–25. https://doi.org/
10.1037/amp0000191
Center for Open Science. (2020). COS: Center for Open Science. https://www.cos.io/
Cohen, J. (1988). Statistical power analysis for the behavioral sciences. Erlbaum.
Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49(12), 997–1003. https://doi.org/
10.1037/0003-066X.49.12.997
Cooper, H. (2007). Evaluating and interpreting research syntheses in adult learning and literacy. National
Center for the Study of Adult Learning and Literacy. https://doi.org/10.1037/e549792010-001
Cooper, H. (2008). The search for meaningful ways to express the effects of interventions. Child
Development Perspectives, 2(3), 181–186. https://doi.org/10.1111/j.1750-8606.2008.00063.x
Cooper, H. (2018). Reporting quantitative research in psychology: How to meet journal article reporting
standards. American Psychological Association. https://doi.org/10.1037/0000103-000
Feibleman, J. K. (1969). An introduction to the philosophy of Charles S. Peirce. MIT Press.
Ferris, T. (2010). The science of liberty. HarperCollins.
Godfrey-Smith, P. (2021). Theory and reality: An introduction to the philosophy of science (2nd ed.).
University of Chicago Press. https://doi.org/10.7208/chicago/9780226771137.001.0001
Hume, D. (1978). A treatise on human nature. Oxford University Press. (Original work published
1739–1740)
Kounios, J., & Beeman, M. (2009). The Aha! moment: The cognitive neuroscience of insight. Current
Directions in Psychological Science, 18(4), 210–216. https://doi.org/10.1111/j.1467-8721.2009.
01638.x
Kuhn, T. S. (1996). The structure of scientific revolutions (3rd ed.). University of Chicago Press.
https://doi.org/10.7208/chicago/9780226458106.001.0001
Maxwell, J. A. (2004). Causal explanation, qualitative research, and scientific inquiry in education.
Educational Researcher, 33(2), 3–11. https://doi.org/10.3102/0013189X033002003
Merton, R. K. (1957). Priorities of scientific discovery. In N. Storer (Ed.), The sociology of science:
Theoretical and empirical investigations (pp. 635–659). University of Chicago Press.
Merton, R. K. (1979). The sociology of science: Theoretical and empirical investigations. University of
Chicago Press.
Murray, S., Boyaci, H., & Kersten, D. (2006). The representation of perceived angular size in human
primary visual cortex. Nature Neuroscience, 9, 429–434. https://doi.org/10.1038/nn1641
Neimark, E. D., & Estes, W. K. (1967). Stimulus sampling theory. Holden-Day.
Rodgers, J. L. (2010). The epistemology of mathematical and statistical modeling: A quiet methodological
revolution. American Psychologist, 65(1), 1–12. https://doi.org/10.1037/a0018326
Rutjens, B. T., Heine, S. J., Sutton, R. M., & van Harrevld, F. (2018). Attitudes toward science. Advances
in Experimental Social Psychology, 57, 125–165. https://doi.org/10.1016/bs.aesp.2017.08.001
Shweder, R. A. (1996). Quanta and qualia: What is the “object” of ethnographic research? In R. Jessor,
A. Colby, & R. A. Shweder (Eds.), Ethnography and human development: Context and meaning is
social inquiry (pp. 175–182). University of Chicago Press.
Webb, E. J., Campbell, D. T., Schwartz, R. D., Sechrest, L., & Grove, J. B. (1999). Unobtrusive measures.
SAGE.
Wilkinson, L., & the Task Force on Statistical Inference. (1999). Statistical methods in psychology
journals: Guidelines and explanations. American Psychologist, 54(8), 594–604. https://doi.org/
10.1037/0003-066X.54.8.594
xlv
2ND PAGES