Escolar Documentos
Profissional Documentos
Cultura Documentos
http://orm.sagepub.com/
A Review and Analysis of the Policy-Capturing Methodology in Organizational Research: Guidelines for
Research and Practice
Ronald J. Karren and Melissa Woodard Barringer
Organizational Research Methods 2002 5: 337
DOI: 10.1177/109442802237115
The online version of this article can be found at:
http://orm.sagepub.com/content/5/4/337
Published by:
http://www.sagepublications.com
On behalf of:
Additional services and information for Organizational Research Methods can be found at:
Email Alerts: http://orm.sagepub.com/cgi/alerts
Subscriptions: http://orm.sagepub.com/subscriptions
Reprints: http://www.sagepub.com/journalsReprints.nav
Permissions: http://www.sagepub.com/journalsPermissions.nav
Citations: http://orm.sagepub.com/content/5/4/337.refs.html
ORGANIZATIONAL
10.1177/109442802237115
Karren,
Barringer / THE
RESEARCH
POLICY-CAPTURING
METHODS METHODOLOGY
Policy-capturing has been employed extensively in the past to examine how organizational decision makers use the information available to them when making
evaluative judgments. The purpose of this article is to provide researchers with
guidelines for enhancing the reliability and validity of their studies. More specifically, the authors identify issues researchers may want to consider when designing
such studies and offer suggestions for effectively addressing them. They draw on a
review of 37 articles from 5 major journals to identify best practice and discuss
the advantages and disadvantages of alternative approaches to resolving the various issues. The key issues are (a) the realism of the approach and its effect on both
internal and external validity, (b) the limits of the full factorial design, (c) the need
for orthogonal cues, (d) sample size and statistical power, and (e) the assessment
of reliability. The analysis also includes comparisons with conjoint analysis, a
similar methodology used in the marketing research literature.
337
338
employee discipline (Klaas & Wheeler, 1990), job analysis (Sanchez & Levine, 1989),
sexual harassment (York, 1989), employment interviews (Dougherty, Ebert, &
Callender, 1986; Graves & Karren, 1992), contract arbitration (Olson, DellOmo, &
Jarley, 1992), and motivation (Zedeck, 1977). Most of these studies have employed an
experimental policy-capturing design. That is, a survey questionnaire is utilized to
elicit judges likely responses to scenarios rather than field data being used to examine
actual responses. A thorough review of the latter approach is available from Roehling
(1993), and we therefore confine our analysis to experimental designs.
The popularity of the policy-capturing method stems from a number of advantages.
First, the method overcomes many of the limitations inherent in other, more direct
approaches to examining individuals decision policies. For example, a simpler
method involves asking individuals to rate or rank the variables of interest in order of
importance (self-report attribute ratings). Concerns about the validity of this method
have been raised by studies finding that such stated decision policies differ from actual
(observed) decision policies (Hitt & Middlemist, 1979; Sherer et al., 1987; Stumpf &
London, 1981). The discrepancy may stem from individuals not being candid in their
responses because of a desire to be socially correct. Rating challenging work above
high pay, for example, may be the socially desirable response to questions about the
importance of job attributes. In fact, studies have shown that less socially desirable factors such as salary receive more weight and are seen as more important when they are
derived through policy-capturing (see Feldman & Arnold, 1978). Policy-capturing
purportedly weakens such social desirability effects by indirectly assessing the importance of explanatory variables and, for this reason, is considered preferable to the selfreport attribute method (Arnold & Feldman, 1981; Judge & Bretz, 1992; Rynes et al.,
1983).1 Moreover, rating or ranking individual attributes requires more self-insight
than making overall judgments about multi-attribute scenarios. Finally, asking individuals to make overall judgments about multi-attribute scenarios is more similar to
actual decision problems, and hence more realistic, than is a self-report attribute
design (Rynes et al., 1983).
Another advantage of the methodology comes from the ability of the researcher to
experimentally manipulate cue values. By minimizing variable intercorrelations,
researchers avoid the problems of multicollinearity often found with field data and
enhance the capacity to assess the independent effects of cues (e.g., Feldman &
Arnold, 1978). Furthermore, policy-capturing is typically carried out at the individual
level. This means that a separate model is generated for each decision maker, although
aggregate analyses of groups of individuals can also be conducted. These separate,
individual analyses allow for a more in-depth assessment of differences between individuals.
These advantages notwithstanding, there are several issues that researchers designing a policy-capturing study must address if they are to avoid questions about the validity of their results. A number of concerns have been raised about the effect of using
simulated decision contexts on the external validity of results. Care must therefore be
taken to create scenarios that include salient and realistically defined cues and to avoid
unlikely cue combinations. Validity may also be compromised by a failure to consider
respondent overload, statistical power, and reliability.
The purpose of this article is to provide researchers interested in using the policycapturing approach with guidelines for enhancing the reliability and validity of their
339
studies. More specifically, we identify the issues researchers may want to consider
when designing a policy-capturing study and offer suggestions for effectively addressing these issues. The aim is to provide those unfamiliar with the methodology with
accessible and comprehensible suggestions for designing their own studies. Our analysis of the issues is informed by a review of the approaches taken by researchers who
have used the policy-capturing design in past research. We also explore how marketing
researchers using conjoint analysis, a methodology for assessing the relative importance of product attributes in consumer decision making, have addressed some of the
issues. Our review of all of these studies is aimed not at critiquing the research designs
but, rather, at developing an understanding of what can be considered best practice.
Where the best practice is unclear, that is, where researchers have used a number of different approaches to address a design issue, we discuss the advantages and disadvantages of each and offer recommendations for putting available techniques to the best use.
We distinguish our article from the Aiman-Smith, Scullen, and Barr study (2002
[this issue]) in two ways. First, Aiman-Smith et al. offer a tutorial that covers a broad
range of topics including study design, execution, analysis, interpretation, and reporting of policy-capturing studies. Our article covers a narrower range of topics but generally goes into more depth on some of the key issues regarding design and execution.
Second, our discussion of these issues incorporates our extensive review of prior studies using the policy-capturing methodology.
We begin our article with a brief review of studies that have used the policy-capturing approach. We have limited our selection of studies to those published over the past
25 years in five highly regarded management journals: the Journal of Applied Psychology, Personnel Psychology, the Journal of Management, the Academy of Management
Journal, and Organizational Behavior and Human Decision Processes. We found 37
studies that used policy-capturing as the primary methodology. These studies are summarized in Table 1. As Table 1 indicates, we found differences across these studies in
the types of decisions made, the samples used, the designs employed, the stimuli presented to the respondents, and the analyses employed by the researchers.
Job choice, ratings of job applicants, and performance evaluation decisions made
up approximately one half of the types of decisions studied. Other popular types were
compensation, disciplinary, and absence decisions. The remaining types of judgments
include promotion, sexual harassment, job task importance, firm acquisition, media
choice, organizational effectiveness, utility, transportation services, and changes in
work attitudes.
The samples for these studies were drawn from a number of different groups.
Undergraduate and graduate students were among the most used subjects in the studies. In most cases, they were used to simulate decisions that would typically involve
them (e.g., job choice). Other studies included decisions made by managers, employees, and specialists such as interviewers and recruiters, executives, and faculty members. Again, in most cases, the decision makers were those who typically had experience with the decisions.
We grouped the designs into four categories: full factorials (16), fractional factorials (4), studies without a factorial design and in which the intercorrelations were zero
or near zero (12), and studies in which the intercorrelations were significantly above or
below zero (4). One study did not report the size of the intercorrrelations. Thus, for
most of the designs, the intercorrelation of the factors was either zero or near zero.
(Text continues on page 344)
340
Table 1
Summary of Published Policy-Capturing Studies
Study by Journal
Decision Context
Sample
47 undergraduates
4 variables (4 3 3 2)
and 55 government
employees
Interviewers ratings
One organization with
8 variables
of job applicants
3 interviewers
Managers evaluations 84 managers from
6 variables: 2 levels each
of job applicants
multiple organizations
Job choice
Evaluating job
applicants and
starting salaries
Changes in work
attributes
Job choice
Estimates of the
dollar value of
performance
Job choice
62 graduate students
from 2 universities
68 line and staff
managers from
multiple organizations
88 department store
salespersons from
one organization
67 undergraduates
from 2 universities
17 managers from
1 organization
10 undergraduates
4 variables (4 3 2 3)
6 applicant attributes:
2 levels each
Stimulus Presentation
Correlation of Variables
36 written
transportation scenarios
36 scenarios
of 72 possible;
intercorrelations < .25
.05 to .97; average
was .52
Less than .10
Slight correlations,
r < .20
Written scenarios
13 variables: multiple
values
50 written profiles of
hypothetical
programmers
Written job descriptions
Orthogonalfull
factorial
Uncorrelated
dimensions
Overall importance
of job tasks
60 employees from
4 jobs in 2 cities
2 compensation
decisions
35 compensation
specialists
(a) 5 factors (2 4
4 5 5); (b) 5 factors
(3 3 5 3 5)
Orthogonalfull
factorial
Correlated (ranged from
.02 to .91)
Near 0; less than .10
Personnel Psychology
Cable & Judge (1994)
Downloaded from orm.sagepub.com at U.A.E University on August 11, 2014
Madden (1981)
5 performance factors:
3 levels each
Written descriptions of
performance
Intercorrelations were
approx. zero
Written descriptions of
variables
Interviewers
evaluations of
job candidates
Disciplinary decisions
of substance abuse
171 undergraduate
5 factors: 2 levels each
students seeking
jobs from 1 university
29 corporate
6 factors: 2 levels each
interviewers from
1 organization
93 managers from
6 factors: 2 levels each
2 organizations
Orthogonalfull
factorial + 4
replicates
Orthogonalfull
factorial
Disciplinary decisions
Written descriptions of
variables
Written profiles
Not reportedprofiles
randomly selected
5 factors: (2 2
3 2 2)
Written profiles of
hypothetical employees
71 Chinese managers
and 218 grad.
student alumni/
managers
4 factors (2 2 2 2)
Written profiles of
employees
Orthogonalfull
factorial + full
replicates
Orthogonalfull
factorial
13 supervisors/10
faculty members
20 faculty members
and chair of
university
Written profiles
Intercorrelations < .2
Written hypothetical
profiles
Intercorrelations were
set to zero
19 human resource
managers and
28 line managers
Performance
3 experiments with
evaluations
undergraduates:
subjects were 58,
70, and 43
Salary raise decisions 11 supervisory
personnel
Compensation
decisions
Written descriptions of
variables
Written case scenarios
Orthogonal; 32 of 64
possible, fractional
design
Orthogonalfull
factorial
341
(continued)
342
Table 1 continued
Downloaded from orm.sagepub.com at U.A.E University on August 11, 2014
Study by Journal
Decision Context
Sample
Stimulus Presentation
Written scenarios
Profiles presented on
cards
Written scenarios
9 performance factors:
3 levels each
Written descriptions of
hypothetical orgs.
Written scenarios
containing behaviors
Written scenarios
Correlation of Variables
Orthogonalfull
factorial
Median correlation = .09
Orthogonalfull
factorial
Intercorrelations were
zero
Intercorrelations were
zero
Orthogonalfull
factorial
25 possible factors:
Simulated cases or
Intercorrelations
5 levels
work units
between criteria,
most below .4
5 factors: 2 levels each
Hypothetical scenarios
Orthogonalfractional
factorial design
6 factors (1/2 2 2
Written acquisition decision Fractional factorial
2 2 2 2)
making exercise
zero correlations
5 factors (2 2 3 2 2) Written hypothetical
Orthogonalfull
candidates
factorial
(a) 5 factors (5 3
Written scenarios
(a) No significant
2 2 2); (b) 3 factors
associations among
(3 2 4)
factors;
(b) orthogonal
full factorial
York (1989)
Journal of Management
Beatty, McCune, &
Beatty (1988)
Policies on sexual
harassment
Written profiles
80 of 384; correlations
were near zero
Compensation
decisions
41 Japanese and 63
U.S. managers from
multiple organizations
65 undergraduate and
graduate students,
2 universities
138 employees in a
university
57 employees or
19 triads (supervisor
and subordinates)
Written descriptions of
hypothetical employees
Average correlation =
.22
Written scenarios
Orthogonalfull
factorial
4 factors (2 2 2 3)
Written scenarios
Written scenarios
Orthogonalfull
factorial
Orthogonalfull
factorial
Job choice
Attributions on the
cause of absence
Absence disciplinary
decisions
343
344
Most of the stimuli presented were written scenarios or profiles. Only a few were
videotaped. Of those written scenarios, almost all were hypothetical situations.
The results of the vast majority of the studies included an assessment of the policies
of individual judges. Among studies that examined both linear and nonlinear effects,
the linear model contained most of the explained variance; thus, little was attributed to
nonlinear or interaction effects. About a third of these studies considered some form of
cluster analysis to group the relative weights. A large minority of the studies also used
between-subjects analyses specifically to test various hypotheses of the study. Some of
the studies included comparisons of the policy-capturing methodology with direct
assessments. The results of these comparisons indicated that the correspondence was
only moderate (e.g., Arnold & Feldman, 1981).
We next discuss the five key issues we have identified as important design considerations for researchers developing a policy-capturing study: (a) the realism of the
approach and its affect on both internal and external validity, (b) the limits of the full
factorial design, (c) the need for orthogonal cues, (d) sample size and statistical power,
and (e) the assessment of reliability. We review how past researchers have addressed
these issues and offer recommendations accordingly.
A recurrent concern about policy-capturing has been with the realism of the decision problems presented to participants and, hence, the external validity of the results.
A realistic decision problem is one that is representative of the problems that occur naturally in the participants environment, whereas an unrealistic problem is one that is
unlikely to occur. If the decision problems used in a policy-capturing study are not
realistic, then the results may be biased and cannot be generalized to nonexperimental
settings (Klaas & Wheeler, 1990; Lane, Murphy, & Marques, 1982; Olson et al., 1992;
York, 1989).
There are a number of inherent challenges to enhancing realism. First, in a policycapturing design, individuals are asked to make judgments based on a limited amount
of information (typically four to eight variables, as shown in Table 1), whereas they are
likely to have more extensive information when making judgments of actual cases
(Olson et al., 1992). There is evidence, however, that owing to cognitive limitations,
individuals tend to base judgments on a relatively small number of criteria (Cooksey,
1996; Rossi & Anderson, 1982; Sanchez & Levine, 1989). Hence, if care is taken to
identify, and include in the study, the decision criteria that are likely to be most salient
to decision makersjudgments, then the realism of the decision problem, and the external validity of the study, can be enhanced. This objective may be difficult to achieve,
however, as it is virtually impossible to ascertain a priori the factors to which individuals attend in making judgments (Viswesvaran & Barrick, 1992; York, 1989). One
approach has been to survey or interview individuals involved in making the decisions
of interest. This may involve focus groups and/or interviews with individuals similar to
the studys sample (Graves & Karren, 1992; Rynes & Lawler, 1983; Viswesvaran &
Barrick, 1992). The advantage of such an approach is that the information is provided
345
by the decision makers themselves, who are arguably in the best position to know what
they consider when making decisions (Cooksey, 1996). A potential disadvantage, on
the other hand, is that the identification of important criteria is a subjective process,
and the information obtained may be incomplete due to lapsed or reconstructed memories (Cooksey, 1996). To minimize these problems, researchers are advised to use multiple, carefully selected decision makers and look for consistent themes (Cooksey,
1996). An alternative approach is to examine company or other related records to
ascertain important decision criteria. Hobson, Mendel, and Gibson (1981) examined
the participating organizationsevaluation policies to identify salient performance criteria. Similarly, York (1989) analyzed 123 related court cases to identify variables for a
study of judgments about sexual harassment complaints. This technique is more
objective than the focus group/interview approach; however, it is limited by the quality
and accuracy of the records (Cooksey, 1996). Still another approach is to review past
theoretical and empirical research (Allen & Muchinsky, 1984; Hollenbeck & Williams, 1987; Pablo, 1994). Where available, such information may provide more
objective support for criterion importance. This approach is likely to be of less use to
researchers exploring new topics. Our recommendation is that researchers obtain
information from all available sources (focus groups, company records, prior
research) and use criteria for which there is consistent support.
Unrealistic decision problems may also be created when a full factorial design is
used to set up decision problems (Klaas & Wheeler, 1990; Lane et al., 1982; York,
1989). This very common policy-capturing design involves creating hypothetical scenarios by completely crossing all of the variables. In this case, the variables are said to
be orthogonal and their independent effects may be accurately assessed. As seen in
Table 1, this full factorial approach has been used by many of the studies reviewed for
this article (see, e.g., Bretz & Judge, 1994; Graves & Karren, 1992; Klaas &
DellOmo, 1991; Rynes et al., 1983).
A potential problem with this approach is that if variables are truly correlated in
the environment, then the decision makers may be presented with unrealistic cases
(York, 1989). Crossing employment status with retirement benefits, for example,
would create one scenario in which a contingent worker received pension benefits.
Two alternative approaches have been developed to address this concern. In one, hypothetical scenarios are created in such a way as to enhance realism and minimize variable intercorrelations. Klaas and Wheeler (1990), for example, used an orthogonal
design but selected factors that are not typically correlated in the real world. Klaas
and DellOmo (1991) provided plausible explanations to study participants for the
seemingly implausible combinations created by orthogonal manipulations. Beatty,
McCune, and Beatty (1988) endeavored to replicate real-world correlations rather
than crossing all explanatory variables; intercorrelations between the eight variables
included in the study were on the average .22. Realism and validity are enhanced using
this approach; however, it is also possible that important criteria would be excluded
because of problems with multicollinearity. A second approach involves using actual,
rather than hypothetical, scenarios. Dougherty et al. (1986), for example, asked participants to rate job applicants after listening to actual interviews that had been tape
recorded. In another study, job incumbents rated the importance of tasks listed in their
(actual) job task inventories (Sanchez & Levine, 1989). Intercorrelations between the
explanatory variables in both of these studies were high. The decision problems are
thus realistic, but because the variables are not orthogonal, assessing their independent
346
effects becomes more difficult. Thus, researchers must in many cases make trade-offs
between realism, on one hand, and acceptable levels of variable intercorrelations, on
the other. In these instances, compromises will be required between the ideally desirable and the practically realizable features of the design (Cooksey, 1996, p. 94). As
to the extent to which researchers should emphasize realism over low
intercorrelations, we note that of all of the studies reviewed for this article, only two
report correlations greater than .5 (Dougherty et al., 1986; Sanchez & Levine, 1989).
Our recommendation is to adopt, wherever possible, the approach taken by the preponderance of studies accepted for publication, that is, ensuring that variable
intercorrelations are 0 or near 0. Higher correlations have in many cases been acceptable and may be unavoidable in situations where the variables of interest are naturally
correlated in the environment. Limited evidence that raw-score regression weights are
similar under different intercorrelation levels (ranging from near 0 to .5) suggests that
0 correlations are not required to estimate variable importance (Lane et al., 1982).
Nevertheless, the impact of higher correlations (i.e., greater than .5) on estimates of
variable importance has not been examined, and our review suggests that levels near 0
or less than .20 are more likely to be favorably received.
The realism of decision problems can also be affected by the operationalization of
explanatory variables. If the manipulated values of the explanatory variables are not
representative of the values observed in the judges environment, the external validity
of the study could be limited (Judge & Bretz, 1992; Rynes et al., 1983). Rynes et al.
(1983) found that differences in the variance in treatment levels can generate different
importance weights. Specifically, the results suggested that the estimated effects of
payrelative to other variables included in the studyare higher when the difference
between defined salary levels is wide than when it is narrow. Range effects were also
obtained in a study conducted by Highhouse, Luong, and Sarkar-Barney (1999),
although these authors reached the somewhat different conclusion that the direction of
these effects may vary across decision contexts. That is, the effects of attribute range
may depend on whether the decision maker is in the initial (prechoice screening) or
ultimate (final) stage of job choice. Both studies suggest, however, that judges
responses to hypothetical scenarios are sensitive to attribute range. Hence, conclusions
about the effects of explanatory variables that are based on responses to hypothetical
scenarios cannot be generalized to actual decision contexts unless treatment levels are
realistically defined.
Ensuring the realism of defined treatment levels, as well as variable combinations,
typically involves obtaining input from individuals familiar with the decision context.
Allen and Muchinsky (1984), for example, asked Department of Transportation officials to review hypothetical scenarios describing proposed transportation services for
the physically handicapped. Staff reviews have also been sought to ensure the realism
of descriptions of hypothetical job candidates (Graves & Karren, 1992), employee disciplinary problems (Klaas & Wheeler, 1990), and substance abuse violations (Klaas &
DellOmo, 1991). Levels of variables examined in studies of students job choices
have been defined based on data from college career offices (Bretz & Judge, 1994;
Cable & Judge, 1994; Judge & Bretz, 1992; Rynes et al., 1983). Field data were also
used to create hypothetical applicant profiles (Mazen, 1990) and acquisitions (Pablo,
1994). Finally, pretests using respondents drawn from the same population as the
study sample have helped researchers to identify and revise unrealistic scenarios
(Pablo, 1994; Webster & Trevino, 1995). These methods have the advantage of
347
As noted above, many of the policy-capturing studies reviewed in this article have
employed a full factorial design in which all variables are completely crossed and bal-
348
anced. Such an approach allows the assessment of the independent effects of each variable on an individual respondents decision (Graham & Cable, 2001; Webster &
Trevino, 1995). A full factorial design also allows the assessment of all main and
higher order effects (Graham & Cable, 2001).
Depending on the number of variables included in the study, however, employing a
full factorial design means that respondents could be asked to review an inordinately
large number of scenarios. Completely crossing seven variables with two levels each,
for example, generates 128 unique scenarios (27 = 128). Employers and/or individuals
are often reluctant to participate in such a time-consuming study, and procuring an
adequate sample under these circumstances can be difficult. Moreover, even among
those individuals willing to read this many scenarios, respondent overload, fatigue,
and stress can affect responses (Graham & Cable, 2001; Webster & Trevino, 1995;
York, 1989). Survey length has been found to be associated with significant differences in respondent stress and exhaustion (Graham & Cable, 2001). Consequently,
researchers may want to limit the number of scenarios presented to participants (Rossi &
Anderson, 1982; Webster & Trevino, 1995). This can be achieved by limiting the number of experimental variables, thereby minimizing the number of scenarios created
when the variables are completely crossed. Limiting the study to a small number of
variables, however, may require the researcher to exclude potentially important
explanatory variables (Graham & Cable, 2001). Hence, researchers employing a full
factorial design may find that they must either limit the scope of their study or risk
compromising the quality of the data (Graham & Cable, 2001).
A common approach to this dilemma is to ask respondents to read a subset of the
full factorial set. This is known as a confounded factorial design, which encompasses
two popular designs: the incomplete block design and the fractional factorial design.
These designs allow the researcher to examine a broader set of variables while avoiding respondent overload (Graham & Cable, 2001). Cue sets can be created whereby
variables are orthogonal, thus allowing the assessment of each variables independent
effects but not higher order (three-way or higher) interaction effects (Cochran & Cox,
1957). If such interactions are not theoretically important, and enough respondents are
available to evaluate all scenarios, then this type of design may be an appropriate
method for reducing respondent boredom and strain (Graham & Cable, 2001; Klaas &
DellOmo, 1991).
Cue sets for the incomplete block design and the fractional factorial design are created in similar fashion. Both involve systematically dividing the full factorial set into
blocks and presenting each respondent with one of the blocks (Cochran & Cox, 1957;
Graham & Cable, 2001; Webster & Trevino, 1995). Each block is composed of a
unique set of scenarios and is created by dividing the full set into halves, quarters,
eighths, and so on (Graham & Cable, 2001).
The major difference between the two types of confounded designs is that all of the
subsets are used (i.e., subgroups of participants each receive a different subset of scenarios) with the incomplete block design, whereas only one subset is used (i.e., all participants receive the same subset of scenarios) with the fractional factorial design.
Consequently, the number of participants required to conduct a study using the incomplete block design increases with the number of blocks used (Graham & Cable, 2001).
That is, at least two participants are required for a one-half (two-block) design, at least
four are needed for a one-quarter (four-block) design, and so on.
349
Owing at least in part to the additional sample and procedural requirements of the
incomplete block design, fractional designs have historically been the approach of
choice among researchers using the confounded factorial approach (Graham & Cable,
2001). Another approach is to randomly select scenarios from the full factorial set,
checking to ensure that variable intercorrelations are not high (Allen & Muchinsky,
1984; Klaas & DellOmo, 1991; Pablo, 1994; Viswesvaran & Barrick, 1992; York,
1989). Allen and Muchinsky (1984), for example, randomly selected 36 out of the 72
possible transportation service descriptions; intercorrelations between the variables
were low (r < .25). Both of these fractional designs allow the researcher to examine the
full range of important cues without creating respondent overload. Because they do
not incorporate the full set of scenarios, however, the researcher cannot be sure that
results are unaffected by the particular set of scenarios selected (Graham & Cable,
2001). Relative to the incomplete block design, which uses the full set of scenarios, the
fractional design allows estimation of fewer effects and requires making more
assumptions about which effects are unimportant (Graham & Cable, 2001).
The confounded factorial design seems to offer researchers a method of studying
how people make decisions without straining respondents or limiting the scope of the
study. Nevertheless, the extent to which this method is a viable alternative to the more
statistically rigorous full factorial design is unclear. We are aware of just one study that
has examined the merits of a full versus a confounded factorial design. In a study of the
effects of five firm attributes on job seekers perceptions of firm reputation, Graham
and Cable (2001) randomly assigned 108 college students to either a full (32 scenarios)
or an incomplete block (8 scenarios) design. Their results indicated that the estimated
effects of explanatory variables were substantially the same across the two designs and
that a full factorial design generated significantly more stress, fatigue, and negative
reactions to survey length among respondents than did an incomplete block design.
Given the limited empirical research on the viability of the confounded factorial
design, it is probably prudent to use a full factorial design wherever possible. One concern is that a design with too few scenarios lacks sufficient power, an issue that is discussed later in the article. Furthermore, scenarios that contain many factors are
cognitively complex and may require some practice with the response scale before
study respondents begin to process the information reliably. In these situations, a confounded factorial design with a small number of scenarios may have insufficient reliability unless respondents are given several practice scenarios before beginning the
task. Nevertheless, the Graham and Cable (2001) study suggested that this type of confounded factorial design is under certain circumstances an acceptable, and perhaps
preferable, alternative. More specifically, these authors argued that where a confounded factorial design is indicated, researchers should give particular consideration
to the incomplete block design. The primary advantage of this approach is that it
reduces the number of scenarios study participants are asked to evaluate. Moreover, as
noted above, estimation of cue effects using the incomplete block design is not limited
by the exclusion of scenarios, as is the case with the fractional design. Hence, when the
number of salient explanatory variables and/or treatment levels included in the study is
such that a full factorial design would generate an inordinately large number of scenarios, researchers should give serious consideration to an incomplete block design.
Determining what constitutes inordinately large may be a matter of judgment at
this point. Rossi and Anderson (1982) suggested an upper limit of 60 scenarios.
350
Aiman-Smith et al. (2002) have recommended using no more than 80 written scenarios. Graham and Cable (2001) found that respondents reacted more negatively to 32
than to 8 scenarios, although the mean stress score for the larger survey was relatively
moderate (3.28 on a 7-point scale). It may be that acceptable survey length varies
across individuals and that pretesting may be needed to determine the maximum number of scenarios respondents can reasonably be expected to process. The optimal number may also vary according to the size of the scenario. Rossi and Anderson, for example, maintained that participants in their study could respond to 60 scenarios in 20
minutes. A smaller number might be indicated where the scenarios require more time
to read and process.
Graham and Cable (2001) also suggested that the incomplete block design may be
most appropriate where the examination of individual decision policies is not a central
focus of the inquiry because individual regression equations must be interpreted with
caution when study participants do not evaluate all possible scenarios. Nevertheless,
we are aware of at least three published studies employing a confounded factorial
design in which data analysis included the estimation of individual regression equations (Klaas & DellOmo, 1991; Pablo, 1994; Webster & Trevino, 1995). According to
Graham and Cable, regression estimates under these circumstances are likely to be
more useful if researchers employ larger fractions (e.g., one half as opposed to one
quarter). Finally, a confounded factorial design is only appropriate where higher order
interactions are expected to be unimportant to explanations of judgments (Graham &
Cable, 2001; Klaas & DellOmo, 1991).
Orthogonality of Cues
351
whether, or under what circumstances, researchers should take care to ensure that variables are completely uncorrelated.
To our knowledge, one study addresses this issue. Lane et al. (1982) compared estimates of variable importance under three different correlation structures and found
that raw-score regression weights did not change across structures, whereas other
measures of importance (simple correlations between explanatory and dependent variables, semipartial correlations, and standardized regression coefficients) did. The
authors concluded that zero intercorrelations are not required to estimate the importance of explanatory variables and that raw-score regression weights are the most
appropriate indicators to use when variables are not orthogonal. They also noted, however, that raw-score regression weights are not independent of the scale used and may
not be ideal in studies where such independence is important. For example, the
observed effect of a one-unit change in intelligence quotient (with a range of 100 or
more) is likely to be considerably smaller than that of a one-unit change in years of
experience (with a range of 25 or less), suggesting, perhaps erroneously, that years of
experience is a more important determinant of the outcome of interest than intelligence. Hence, in decision problems where the variance of cues is dissimilar, the relative importance of these cues cannot be accurately assessed unless the regression
weights are standardized (i.e., independent of the scale). In such cases, Lane et al. recommended using standardized regression coefficients, as the results of their study suggest that the change in these estimates across different correlation structures, although
significant, is relatively small (p. 238).
Variable intercorrelations in the Lane et al. (1982) study were not very high, so it is
not clear whether estimates of variable importance when intercorrelations are high can
be interpreted with any confidence. Furthermore, use of a nonorthogonal design may
limit the researchers choice of importance measures and make the use of cluster analysis more difficult. Zero or near-zero variable intercorrelation structures would therefore seem to be the preferred design. If an orthogonal design is not possible (due, perhaps, to the creation of unrealistic scenarios), the researcher may want to consider a
design in which variable intercorrelations are relatively low and raw-score regression
weights are used as the measure of importance.
Sample Size and Power
One way to increase the power of a research study is to increase the number of subjects or participants in the study (Cohen, 1988). When planning a research study,
researchers will try to reduce Type II errors and increase power. Specifically, researchers prefer having a reasonably large chance of rejecting the null hypothesis if it, in fact,
is false. However, when designing a policy-capturing study, the number of subjects
may take a secondary role because the main focus typically is the individual analysis
for each subject. In this case, the power of the individual analysis is not based on the
number of subjects, because there is only one subject, but the number of scenarios or
judgments made by each subject. The number of scenarios determines the number of
observations for each analysis. That is, whether the regression weights will be significant is likely to be related to the number of scenarios. In a study employing multiple
regression techniques, the preferred ratio of the number of scenarios to factors is 10:1,
but the minimum ratio is considered to be 5:1 (Cooksey, 1996). These ratios are guide-
352
lines, and other factors should be considered, such as the extent of the explained variation in judgments and the intercorrelations of the cues (see Cooksey, 1996, for further
discussion). However, as discussed above, increasing the number of factors, and
hence the number of scenarios, may also create problems such as stress and exhaustion (Graham & Cable, 2001). Thus, there seems to be a trade-off associated with a
more comprehensive survey. Although it offers higher power, a comprehensive survey
may result in fatigue and, furthermore, the likelihood of reduced reliability.
As discussed previously, researchers sometimes use a confounded factorial design
rather than a full factorial design to reduce fatigue. In some cases, this is done so the
researcher can add additional cues into the judgment process without burdening the
respondent with a large number of scenarios. That is, there are situations or contexts in
which the decision maker generally considers a large number of cues before making a
judgment, creating the need for fractionalization. However, fractionalization also
results in fewer scenarios, and this, in turn, reduces the power of the design. Thus, it is
important that researchers consider both power and fatigue when determining the
number of scenarios to include in the final design. One approach may be to develop at
least five scenarios for each cue. This would be the minimum ratio for studies employing regression techniques (Cooksey, 1996).
Although the size of the sample does not affect the power of the individual analysis,
it can be an issue in studies using other types of analysis. For example, if the researcher
is going to cluster subjects by strategy, large sample sizes offer a more comprehensive
analysis of the grouping process. Large sample sizes are also likely to be more effective when analyzing individual differences between the respondents. For example, in
the Graves and Karren (1992) study, 29 interviewers were used to analyze the different
decision-making strategies. After doing a cluster analysis, 13 clusters were found to
differentiate the various decision-making strategies among the interviewers. Although
it was likely that most if not all of the decision strategies were found, it was rather difficult to estimate the relative popularity of each cluster because there were 13 clusters
among 29 interviewers. Larger sample sizes allow better estimation of the relative
popularity of the clusters. For instance, Klaas and DellOmo (1991) used a much larger
sample size (93 managers). Their cluster analysis indicated 7 clusters, and with their
much larger sample size, they were able to estimate the popularity of each cluster.
Relatively small (less than 50 subjects) samples are not uncommon among the published studies using the policy-capturing approach. Rynes et al. (1983), for example,
examined the job application decisions made by 10 college seniors. Because there
were only 10 subjects, individual analyses were conducted, but subjects were not clustered. Having few subjects makes it difficult to do any form of clustering. Furthermore,
it is unlikely that any kind of generalization can realistically be made regarding these
students, as small samples may not be representative of the population. Obtaining representative samples that can be used to generalize results to the population is not possible
without sufficiently large numbers of respondents and the use of probability-sampling
techniques.
The smallest sample among the published studies reviewed for this article consisted
of three subjects (Dougherty et al., 1986). However, in this case, the objective was not
to discern the relative effects of each of the cues but to determine the validity of each of
the three decision makers in making decisions about job candidates. Studies that use
relatively few subjects probably have very different objectives; they are unlikely to
make inferences regarding the external validity of their results. Thus, large sample
353
sizes should not be a requirement for all policy-capturing studies. It is necessary that
the researcher desire sufficient power when testing hypotheses.
Reliability
Our analysis of key design issues has so far relied on a review of policy-capturing
studies in the management literature. A number of the issues we discuss are also
addressed by Aiman-Smith et al. (2002). Both articles consider the issue of realism in
the design and presentation of scenarios to participants. Our article considers a variety
of approaches and designs that may be utilized to create realism. We discuss our preference for zero or near-zero intercorrelations between the cues when designing a
study, then consider the advantages and disadvantages of various alternative designs
(e.g., fractional and incomplete block design) to the full factorial design. Furthermore,
in the next section, we discuss approaches used by conjoint analysis researchers to deal
with the realism problem. We also propose that realism is more important than consistency when considering the range or difference between cue levels (values). This
354
Table 2
A Comparison of Conjoint Analysis and Policy-Capturing
Conjoint Analysis
Type of analysis
Cue presentation method
Survey design
Level of analysis
Aggregate analyses
Evaluation of stimuli
Decompositional
Profiles method, trade-off
method, pairwise combination,
adaptive designs
Full factorial design, fractional
design
Individual
Yes, including clustering
Metric scales, nonmetric
procedures
Policy-Capturing
Decompositional
Profiles method
means that in some cases, some cues will have wider ranges than others. Aiman-Smith
et al., on the other hand, propose that ranges should be about the same even if cue levels
are less realistic. We are in agreement, however, about the importance of consistency
with the number of levels across cues. We also discuss the advantages of orthogonality
among cues and the use of raw regression weights and standardized weights when
there are nonorthogonal designs.
Both articles also address the issue of fatigue and the cognitive limits of the decision
maker. Aiman-Smith et al. (2002) advise the researcher not to use more than five cues.
They also state that if written scenarios are used, the total number of scenarios should
be between 50 and 80. We tend to be less prescriptive. Because there are many potential designs and approaches, we do not specify the number of cues or scenarios to use.
We believe, however, that there should be an absolute minimum ratio of scenarios to
cues (i.e., 5:1).
Finally, both we and Aiman-Smith et al. (2002) address the reliability issue and
suggest that more estimates are desirable. Both advise using duplicate scenarios to calculate reliability. We suggest that these duplicates may assist in lessening start-up
effects, a problem found, as discussed by Aiman-Smith et al., when subjects initially
learn the task.
We next consider the insights offered by studies from the marketing literature that
have utilized a very similar methodologyconjoint analysis.
Conjoint Analysis
Conjoint analysis is a methodology that has been used extensively in marketing and
consumer research to understand how consumers evaluate preferences for products or
services (Green & Srinivasan, 1978, 1990). Like policy-capturing, conjoint analysis
uses an individualized factorial survey approach to examine the effects of product or
service attributes on evaluative judgments. An examination of this methodology may
therefore yield useful guidelines for researchers employing a policy-capturing
approach. We begin with a consideration of the methodological and computational
similarities between the two approaches. Our comparison of these approaches is summarized in Table 2. We include in this discussion an examination of the approaches
market researchers have taken to resolve some of the issues (e.g., information over-
355
356
of respondent fatigue often associated with other presentation methods (Hair et al.,
1998).
As with policy-capturing studies, the problems associated with information overload are also concerns in the design of conjoint-measurement studies. Market
researchers have addressed these concerns in much the same way as researchers using
the policy-capturing approachby employing a fractional factorial design. Using a
fractional factorial design serves to reduce the number of scenarios to a manageable
size while maintaining orthogonality (Green, 1974). As discussed earlier, these
designs result in a reduction of interpretable interaction effects. This is not a problem,
however, if the decision maker is using an additive model, as most of the interpretable
variance is assigned to the main effects.
Researchers using the conjoint-analysis method have examined the effects of as
many as 30 factors and have consequently developed a number of alternative
approaches to the fractional factorial design not found in the policy-capturing literature. The hybrid or adaptive approaches, for example, involve a two-stage procedure in
which respondents are first asked to rate the desirability of the full set of factors.2 The
adaptive approach is the more popular of these methods, perhaps because software is
available that allows the researcher to generate individualized scenarios and is probably most useful when examining 10 or more factors (Green & Srinivasan, 1990). In this
approach, each respondent receives a set of scenarios that only include those factors
designated as the most important in the first stage of the procedure. The scenarios are
then evaluated in the same way as the above methods (Hair et al., 1998).
A third issue for researchers using either the conjoint-analysis or policy-capturing
approach is the impact of forced orthogonality on external validity. That is, creating
a design in which the variables are orthogonal but are naturally correlated in the environment may produce profiles that are not representative of the environment familiar
to the respondents (Green & Srinvasan, 1978, 1990). Where variables are substantially
correlated, Green and Srinvasan (1978) have suggested the use of composite factors,
which provide a summary measure of all correlated subfactors. For example, a medical-cost-sharing variable would summarize the overall level of deductible and
coinsurance provisions, which tend to be highly correlated across health care plans.
This approach avoids the problem of creating unrealistic profiles (e.g., high deductible
and low coinsurance); however, it does not allow the researcher to partial out the
effects of subfactors that may be of more interest to the study.
Alternatively, Steckel, DeSarbo, and Mahajan (1991) devised a new optimizing
methodology that entails creating a survey to ensure that variables are as orthogonal
as possible. A combinatorial optimization procedure is used to create a modified fractional factorial design by identifying and excluding nonrepresentative or unrealistic
profiles (i.e., very unlikely to occur in the environment). An algorithm finds a subset of
the realistic profiles that are as close as possible to being orthogonal. This is not unlike
some policy-capturing designs wherein the researcher takes care to create stimulus
sets in which variable intercorrelations are minimized and realism is enhanced (Beatty
et al., 1988; Klaas & DellOmo, 1991; Klaas & Wheeler, 1990).
Data Analysis
Similarities may also be found in the computational issues that arise in conjointanalysis and policy-capturing studies. One issue is the level of analysis. In both
357
approaches, analyses are typically carried out at the individual level, which means that
the analyst generates a separate model for predicting preferences for each respondent.
At the individual level, each respondent rates enough profiles that the analyses can be
performed separately for each person. Predictive accuracy is calculated for each
respondent rather than for the whole sample. One of the common uses of these analyses is to group individuals with similar importance values into segments. Researchers
using either conjoint analysis or policy-capturing are interested in better understanding these segments and may combine this information with other variables such as
demographics to derive respondent groupings that are similar in preference (Hair et al.,
1998). Many times, however, an aggregate analysis is used to estimate the relative
importance of the attributes of the whole sample. In some studies, this between-group
analysis was used to test hypotheses about the average effects of attributes.
A second computational issue involves the specification of the respondents composition rule. The rule describes how the respondent combines the factors to obtain
overall worth. The most common rule invokes the additive model, which assumes that
the respondent simply adds up the values of each attribute to get a total value for the
factor or attribute combination. As is the case in policy-capturing studies, the main
effects account for most of the total variation in preferences, and hence, this model suffices for most consumer applications. Alternatively, the composition rule using interactive effects allows for the interaction of two or more attributes. The choice of a composition rule determines the types and number of stimuli that the decision maker must
evaluate. More stimuli are required if the researcher is interested in evaluating interactions. Consider a study using 4 factors and 4 levels. If the factors are presented using a
full-profile method, in which all factors are included in all scenarios, and the
researcher is only interested in estimating main effects, then just 16 (4 factors 4 levels) of the 256 (44) possible scenarios are needed (Hair et al., 1998). That is, it is possible to estimate main effects using a fractional factorial design. However, the 16 scenarios must be carefully constructed for orthogonality to arrive at the correct estimation of
the main effects. If, on the other hand, interactions are specified as important, additional scenarios are required to assess these effects. In this case, a full factorial design,
with the full set of 256 scenarios (240 more than are needed to assess main effects)
would be required to assess the importance of all 11 interactions.
Research Results
358
Conclusion
Our analysis of some of the key issues related to the policy-capturing method has
indicated not only that this method has been a very effective tool in understanding the
processes by which individual decision makers integrate cues of information in making various types of organizational decisions but that the method is highly flexible and
able to adapt to many contexts. Our comparison with conjoint analysis has been quite
beneficial, as research advances to this methodology can be of assistance in solving
some of the key methodological problems of the policy-capturing method. Although
we are both positive and quite hopeful regarding the scope and versatility of future
research and practice, the researcher designing a study using this method should
understand its limitations and constraints.
To design a sound and valid policy-capturing study, the researcher should first focus
on the purpose of the research. Clearly, the researchers intentions are critical in determining how to deal with problematic situations when no ideal solution is available. In
our analysis, we encountered a number of clear trade-offs when evaluating the various
policy-capturing studies. For example, if researchers are going to use an experimental
design, they typically have to choose between a full factorial design and a confounded
factorial design. In the former, the obvious advantages are the ability to assess the full
model, both the linear and nonlinear components, and to find the relative contribution
of all the variables. On the other hand, the investigator has to consider problems of
stress and fatigue, especially if there are many variables, and also whether the additional scenarios will result in unreliable judgments. Using a confounded factorial
design, the researcher may want to consider the loss of power with fewer scenarios and
the limitations related to the inability of assessing higher order interactions. These
considerations, however, may become moot if and when the required number of cues is
far too large to use a full factorial design. In situations where there are far too many
cues, researchers are likely to utilize statistical packages, which create scenarios with
near-zero correlations between the cues. In cases where there are more than 10 factors,
researchers should utilize the hybrid and adaptive techniques that conjoint researchers
have used for many years.
Design decisions by researchers are not all based on some form of trade-off. We
believe that researchers should be careful to avoid designing policy-capturing studies
that lack realism and result in poor validity. In planning their studies, they may want to
check their variables carefully to avoid cues that are correlated in real situations and
would result in implausible combinations if forced to be uncorrelated in the experiment. They should ensure that levels of each cue are applicable to real settings. Fur-
359
thermore, they may check to see if the constructed scenarios are likely to be interpreted
appropriately within the context of the study. Once this is known, they can then take the
necessary steps to enhance realism. Finally, most of the past studies have not included
a reliability check. Because reliability is a necessary condition for validity, some form
of replication should be included as part of the policy-capturing study.
Notes
1. Evidence suggests that the more sophisticated judges see through the indirect approach,
and eliminating social desirability effects altogether may not always be possible (Mazen, 1990).
2. This so-called self-explication procedure is similar to an approach used in policy-capturing
studies in which researchers run focus groups or interview individuals familiar with the decision
problems to identify salient decision criteria. In policy-capturing studies, factors are invariant
across respondents, whereas in these alternative methods, the factors vary according to individual respondents desirability ratings.
References
Aiman-Smith, L., Scullen, S. E., & Barr, S. H. (2002). Conducting studies of decision making in
organizational contexts: A tutorial for policy-capturing and other regression-based techniques. Organizational Research Methods, 5, 388-414.
Allen, J. S., & Muchinsky, P. M. (1984). Assessing raters policies in evaluating proposed services for transporting the physical handicapped. Journal of Applied Psychology, 69, 3-11.
Arnold, H. J., & Feldman, D. C. (1981). Social desirability response bias in self-report choice
situations. Academy of Management Journal, 24, 377-385.
Beatty, J. R., McCune, J. T., & Beatty, R. W. (1988). A policy-capturing approach to the study of
United States and Japanese managers compensation decisions. Journal of Management,
14, 465-474.
Brannick, M. T., & Brannick, J. P. (1989). Nonlinear and noncompensatory processes in performance evaluation. Organizational Behavior and Decision Processes, 44, 97-122.
Bretz, R. D., Jr., & Judge, T. A. (1994). The role of human resource systems in job applicant decision processes. Journal of Management, 20, 531-551.
Cable, D. M., & Judge, T. A. (1994). Pay preferences and job search decisions: A person-organization fit perspective. Personnel Psychology, 47, 317-348.
Carmines, E. G., & Zeller, R. A. (1979). Reliability and validity assessment. Beverly Hills, CA:
Sage.
Cochran, W. G., & Cox, G. M. (1957). Experimental designs. New York: John Wiley.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Hillsdale, NJ:
Lawrence Erlbaum.
Cooksey, R. W. (1996). Judgment analysis: Theory, methods, and applications. San Diego, CA:
Academic Press.
Cooper, W. H., & Richardson, A. J. (1986). Unfair comparisons. Journal of Applied Psychology,
71, 179-184.
Darlington, R. B. (1968). Multiple regression in psychological research and practice. Psychological Bulletin, 69, 161-182.
Dougherty, T. W., Ebert, R. J., & Callender, J. C. (1986). Policy capturing in the employment interview. Journal of Applied Psychology, 71, 9-15.
Dunn, W. S., Mount, M. K., Barrick, M. R., & Ones, D. S. (1995). Relative importance of personality and general mental ability in managers judgments of applicant qualifications.
Journal of Applied Psychology, 80, 500-509.
360
Feldman, D. C., & Arnold, H. J. (1978). Position choice: Comparing the importance of organizational and job factors. Journal of Applied Psychology, 63, 706-710.
Graham, M. E., & Cable, D. M. (2001). Consideration of the incomplete block design for policycapturing research. Organizational Research Methods, 4, 26-45.
Graves, L. M., & Karren, R. J. (1992). Interviewer decision processes and effectiveness: An experimental policy-capturing investigation. Personnel Psychology, 45, 313-340.
Green, P. E. (1974). On the design of choice experiments involving multifactor alternatives.
Journal of Consumer Research, 1, 61-68.
Green, P. E., & Srinivasan, V. (1978). Conjoint analysis in consumer research: Issues and outlook. Journal of Consumer Research, 5, 103-123.
Green, P. E., & Srinivasan, V. (1990). Conjoint analysis in marketing: New developments with
implications for research and practice. Journal of Marketing, 54(4), 3-19.
Hair, J. F., Jr., Anderson, R. E., Tatham, R. L., & Black, W. C. (1998). Multivariate data analysis
(5th ed.). Englewood Cliffs, NJ: Prentice Hall.
Highhouse, S., Luong, A., & Sarkar-Barney, S. (1999). Research design, measurement, and effects of attribute range on job choice: More than meets the eye. Organizational Research
Methods, 2, 37-48.
Hitt, M. A., & Barr, S. H. (1989). Managerial selection decision models: Examination of
configural cue processing. Journal of Applied Psychology, 74, 53-61.
Hitt, M. A., & Middlemist, R. D. (1979). A methodology to develop the criteria and criteria
weightings for assessing subunit effectiveness in organizations. Academy of Management
Journal, 22, 356-374.
Hobson, C. J., Mendel, R. M., & Gibson, F. W. (1981). Clarifying performance appraisal criteria. Organizational Behavior and Human Performance, 28, 164-188.
Hollenbeck, J. R., & Williams, C. R. (1987). Goal importance, self-focus, and the goal-setting
process. Journal of Applied Psychology, 72, 204-211.
Judge, T. A., & Bretz, R. D., Jr. (1992). Effects of work values on job choice decisions. Journal
of Applied Psychology, 77, 261-271.
Judge, T. A., & Martocchio, J. J. (1996). Dispositional influences on attributions concerning absenteeism. Journal of Management, 22, 837-861.
Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk.
Econometrica, 47, 263-289.
Kennedy, P. (1989). A guide to econometrics (2nd ed.). Cambridge, MA: MIT Press.
Klaas, B. S., & DellOmo, G. G. (1991). The determinants of disciplinary decisions: The case of
employee drug use. Personnel Psychology, 44, 813-835.
Klaas, B. S., & Wheeler, H. N. (1990). Managerial decision making about employee discipline:
A policy-capturing approach. Personnel Psychology, 43, 117-134.
Lane, D. M., Murphy, K. R., & Marques, T. E. (1982). Measuring the importance of cues in policy capturing. Organizational Behavior and Human Performance, 30, 231-240.
Madden, J. M. (1981). Using policy-capturing to measure attitudes in organizational diagnosis.
Personnel Psychology, 34, 341-350.
Martocchio, J. J., & Judge, T. A. (1994). A policy-capturing approach to individuals decisions
to be absent. Organizational Behavior and Human Decision Processes, 57, 358-386.
Martocchio, J. J., & Judge, T. A. (1995). When we dont see eye to eye: Discrepancies between
supervisors and subordinates in absence disciplinary decisions. Journal of Management,
21, 251-278.
Mazen, A. M. (1990). The moderating role of social desirability, age, and experience in human
judgment: Just how indirect is policy capturing? Organizational Behavior and Human Decision Processes, 45, 19-40.
Olson, C. A., DellOmo, G. G., & Jarley, P. (1992). A comparison of interest arbitrator decisionmaking in experimental and field settings. Industrial and Labor Relations Review, 45, 711723.
361
Orr, J. M., Sackett, P. R., & Mercer, M. (1989). The role of prescribed and nonprescribed behaviors
in estimating the dollar value of performance. Journal of Applied Psychology, 74, 34-40.
Pablo, A. L. (1994). Determinants of acquisition integration level: A decision-making perspective. Academy of Management Journal, 37, 803-836.
Pedhazur, E. J. (1982). Multiple regression in behavioral research. New York: Holt, Rinehart
and Winston.
Roehling, M. V. (1993). Extracting policy from judicial opinions: The dangers of policy capturing in a field setting. Personnel Psychology, 46, 477-502.
Rossi, P. H., & Anderson, A. B. (1982). The factorial survey approach: An introduction. In P. H.
Rossi & S. L. Nock (Eds.), Measuring social judgments: A factorial survey approach.
Beverly Hills, CA: Sage.
Rynes, S., & Lawler, J. (1983). A policy-capturing investigation of the role of expectancies in
decisions to pursue job alternatives. Journal of Applied Psychology, 68, 620-631.
Rynes, S. L., Schwab, D. P., & Heneman, H. G., III. (1983). The role of pay and market pay variability in job applicant decisions. Organizational Behavior and Human Performance, 31,
353-364.
Sanchez, J. I., & Levine, E. L. (1989). Determining important tasks within jobs: A policycapturing approach. Journal of Applied Psychology, 74, 336-342.
Sherer, P. D., Schwab, D. P., & Heneman, H. G., III. (1987). Managerial salary-raise decisions:
A policy-capturing approach. Personnel Psychology, 40, 27-38.
Stahl, M. J., & Zimmerer, T. W. (1984). Modeling strategic acquisition policies: A simulation of
executives acquisition decisions. Academy of Management Journal, 27, 369-383.
Steckel, J., DeSarbo, W. S., & Mahajan, V. (1991). On the creation of acceptable conjoint analysis experimental designs. Decision Sciences, 22, 435-442.
Stumpf, S. A., & London, M. (1981). Capturing rater policies in evaluating candidates for promotion. Academy of Management Journal, 24, 752-766.
Viswesvaran, C., & Barrick, M. R. (1992). Decision-making effects on compensation wages:
Implications for market wages. Journal of Applied Psychology, 77, 588-597.
Webster, J., & Trevino, L. K. (1995). Rational and social theories as complementary explanations of communication media choices: Two policy-capturing studies. Academy of Management Journal, 38, 1544-1572.
Wittink, D. R., Krishnamurthi, L., & Nutter, J. B. (1982). Comparing derived importance
weights across attributes. Journal of Consumer Research, 8, 471-474.
Wittink, D. R., Krishnamurthi, L., & Reibstein, D. J. (1990). The effect of differences in the
number of attribute levels on conjoint results. Marketing Letters, 1, 113-129.
York, K. M. (1989). Defining sexual harassment in workplaces: A policy-capturing approach.
Academy of Management Journal, 32, 830-850.
Zedeck, S. (1977). An information processing model and approach to the study of motivation.
Organizational Behavior and Human Performance, 18, 47-77.
Zedeck, S., & Cascio, W. F. (1982). Performance appraisal decisions as a function of rater training and purpose of the appraisal. Journal of Applied Psychology, 67, 752-758.
Zedeck, S., & Kafry, D. (1977). Capturing rater policies for processing evaluation data. Organizational Behavior and Human Performance, 18, 269-294.
Zhou, J., & Martocchio, J. J. (2001). Chinese and American managers compensation award decisions: A comparative policy-capturing study. Personnel Psychology, 54, 115-145.
Ronald J. Karren is an associate professor of management at the University of Massachusetts at Amherst.
He obtained his Ph.D. in Psychology from the University of Maryland. His current research interests include
testing, interviewing, and human resource decision making.
Melissa Woodard Barringer is an associate professor of management at the University of Massachusetts at
Amherst. She received a Ph.D. in Industrial and Labor Relations from Cornell University. Her research interests are in the areas of total compensation and alternative work arrangements.