Você está na página 1de 12

See

discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/225476807

What Reviewers Should Expect from Authors


Regarding Common Method Bias in
Organizational Research

Article in Journal of Business and Psychology · September 2010


DOI: 10.1007/s10869-010-9181-6

CITATIONS READS

764 1,950

2 authors:

Jim Conway Chuck Lance


Central Connecticut State University University of the Western Cape
38 PUBLICATIONS 3,362 CITATIONS 113 PUBLICATIONS 9,314 CITATIONS

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

SMMULs View project

All content following this page was uploaded by Chuck Lance on 12 December 2015.

The user has requested enhancement of the downloaded file.


J Bus Psychol (2010) 25:325–334
DOI 10.1007/s10869-010-9181-6

What Reviewers Should Expect from Authors Regarding


Common Method Bias in Organizational Research
James M. Conway • Charles E. Lance

Published online: 22 May 2010


 Springer Science+Business Media, LLC 2010

Abstract We believe that journal reviewers (as well as are often raised during the review process. While there is a
editors and dissertation or thesis committee members) have substantial literature about controlling common method
to some extent perpetuated misconceptions about common bias that targets researchers and authors (e.g., Cote and
method bias in self-report measures, including (a) that Buckley 1987; Crampton and Wagner 1994; Lance et al., in
relationships between self-reported variables are necessarily press; Podsakoff et al. 2003; Richardson et al. 2009;
and routinely upwardly biased, (b) other-reports (or other Spector 2006; Williams et al. 1989) there is very little that
methods) are superior to self-reports, and (c) rating sources advises reviewers (one exception is Richardson et al.
(e.g., self, other) constitute measurement methods. We argue 2009). This is important because there are misconceptions
against these misconceptions and make recommendations about common method bias (as we explain later) that
for what reviewers (and others) should reasonably expect reviewers, dissertation committees, editors, and other
from authors regarding common method bias. We believe it ‘‘gatekeepers’’ play an important role in perpetuating
is reasonable to expect (a) an argument for why self-reports (Lance and Vandenberg 2009). We intend to fill this gap in
are appropriate, (b) construct validity evidence, (c) lack of the literature by describing misconceptions regarding
overlap in items for different constructs, and (d) evidence common method bias and what it is reasonable for
that authors took proactive design steps to mitigate threats of reviewers to expect from authors to minimize common
method effects. We specifically do not recommend post hoc method bias. By extension, we hope to help authors be
statistical control strategies; while some statistical strategies better prepared to (a) design studies that minimize common
are promising, all have significant drawbacks and some have method bias and (b) present their research in a way that
shown poor empirical results. proactively addresses concerns with method bias.
The literature on common method bias has dealt with a
Keywords Common method bias  Method variance  variety of types of alleged measurement methods including
Self-report measures  Reviewing self-reports (e.g., self-ratings of job satisfaction and work
behaviors), rater effects (e.g., performance ratings by
multiple sources), and assessment center exercises. Our
The focus of this article is on common method bias from a main focus will be on the use of self-reports as measure-
journal reviewer’s perspective. According to Chan (2009) ment method because:
and Spector (2006), concerns about common method bias
(a) It is widely assumed that common method bias
inflates relationships between variables measured by
J. M. Conway (&) self-reports. For example, Podsakoff and Todor
Department of Psychology, Central Connecticut State (1985) stated ‘‘Invariably, when self-report measures
University, 1615 Stanley Street, New Britain, CT 06050, USA
e-mail: conwayj@ccsu.edu obtained from the same sample are utilized in
research, concern over same-source bias or general
C. E. Lance method variance arises’’ (p. 65). A second example
University of Georgia, Athens, GA, USA comes from Organ and Ryan’s (1995) meta-analysis
e-mail: clance@uga.edu

123
326 J Bus Psychol (2010) 25:325–334

of correlates of organizational citizenship behavior variables are necessarily and routinely upwardly biased,
(OCB) in which they stated that ‘‘studies that use self- (b) that other-reports (or other methods) are superior to
ratings of OCB along with self-reports of disposi- self-reports, and (c) that rating sources (e.g., self, other)
tional and attitudinal variables invite spuriously high constitute mere alternative measurement methods.
correlations confounded by common method vari-
ance’’ (p. 779). As a final example, and as part of his Misconception #1: Relationships Between
commentary as outgoing editor of the Journal of Self-Reported Variables are Routinely Upwardly
Applied Psychology Campbell (1982) wrote: Biased
With perhaps one or two exceptions there has
This is the most fundamental misconception. An example
been very little opportunity to exercise any
of this belief in the published literature comes from Organ
professional biases (I think). One possible
and Ryan’s (1995) meta-analysis of correlates of OCB.
exception pertains to the use of a self-report
Organ and Ryan stated that ‘‘The most notable moderator
questionnaire to measure all the variables in a
of these correlations appears to be the use of self- versus
study. If there is no evident construct validity for
other-rating of OCB; self-ratings are associated with higher
the questionnaire measure or no variables that
correlations, suggesting spurious inflation due to common
are measured independently of the question-
method variance’’ (p. 775). Spector (2006) provided indi-
naire, I am biased against the study and believe
rect evidence against inflation, observing that correlations
that it contributes very little. Many people share
among self-report variables are at least sometimes near-
this bias. (p. 692).
zero, and that sometimes different-method correlations are
(b) As a result of this apparently widely held belief,
higher than same-method correlations. However, there is a
authors often report reviewer and editorial concerns
degree of truth to this belief—common (or correlated)
about inflation in self-report relationships (Brannick
method effects can upwardly bias relationships—but the
et al., in press; Spector 2006). Given the tendency for
situation is more complicated by several issues, one of
reviewer criticisms to center on self-reported vari-
which is the important distinction between method vari-
ables, this will be the main focus of our article. We
ance and common methods bias—method variance is
will not emphasize alleged ‘‘method’’ effects due to
systematic error variance due to the method of measure-
rating sources (e.g., multisource performance ratings),
ment and common method bias is inflation of relationships
assessment center exercises or measurement occa-
by shared method variance.
sions because as we explain later, it is a misconcep-
Under a simple extension of classical test theory (Lord
tion that these measurement facets should necessarily
and Novick 1968), the measurement equation for an
be thought of as simply different methods of
observed score (Xijk) represents the kth realization of the ith
measurement (Lance et al. 2009).
individual’s true score on trait (construct) T, as measured
We believe that there are widely held misconceptions by the jth measurement method (Mj) and non-systematic
about the effects of common method bias in self-report measurement error (Eijk):
variables and that reviewers sometimes base criticisms of Xijk ¼ kTij Ti þ kMij Mj þ Eijk ð1Þ
manuscripts on these misconceptions. We will therefore
include a section dealing with misconceptions. This is not Under the traditional assumptions that Ti and Mj are
to say that common method bias is never a problem. There uncorrelated with Eijk and each other (e.g., Widaman
are situations in which common method bias should be an 1985), and that the Xijk, Ti, and Mj have unit variances (i.e.,
important concern, and we will follow the discussion of that the ks are standardized regression coefficients or factor
misconceptions with a section in which we lay out the loadings), the variance in observed scores, r2Xijk ; reflects
kinds of things reviewers should look for and reasonably three independent components: individuals’ true score
expect from authors to minimize common method bias. variability, k2Tij (this can also be interpreted as a squared
reliability index), variance attributable to the method of
measurement, k2Mij , and non-systematic error score
Misconceptions About Common Method Bias—What variance, r2Eijk :
Should Reviewers Not Assume r2Xijk ¼ k2Tij þ k2Mij þ r2Eijk ð2Þ

We suggest that there are three prevailing misconceptions It is in this sense that method variance is interjected into
about common method bias that can impede the progress observed scores and the primary implication is compro-
of research: (a) that relationships between self-reported mised construct validity to the extent that method variance

123
J Bus Psychol (2010) 25:325–334 327

dominates over true score variance. That is, to the extent attenuating effects of measurement error offset (in fact,
that k2Mij dominates over k2Tij in Eq. 2, X reflects less of the were somewhat larger than) the inflationary effects of
intended construct than it does the method employed to common method bias, (b) that same-method observed score
measure it. But compromised construct validity does not correlations are actually quite accurate representations of
necessarily lead to common method bias, or upward dis- their true-score counterparts, and (c) the widespread belief
tortion in measured relationships. In fact, if two traits (i.e., that common method bias serves to inflate common method
TX and TY) are measured by two different, uncorrelated correlations as compared to their true-score counterparts is
methods, the expected correlation will be attenuated as substantially a myth.
compared to the no-method-variance situation. This is
because each measure will contain a source of variance
(due to the measurement method) that is unshared with the Misconception #2: Other-Reports (or Other Methods)
other measure, decreasing the resulting correlation value are Superior to Self-Reports
(the appendix provides a mathematical description of why
the attenuation occurs). A corollary to misconception #1 is that other-reports or
However, inflationary common method bias can occur if other methods (e.g., objective measures) are superior to
different traits are measured by the same method. In the self-reports. This idea is illustrated by the earlier quotations
case in which X and Y are measured by the same method M, from Campbell (1982), Organ and Ryan (1995), and Pod-
the observed correlation can be represented as: sakoff and Todor (1985). These quotations imply that
rXY ¼ kXTX kYTY qTX TY þ kXM kYM ð3Þ including a non-self-reported variable can improve the
quality of the study and/or reduce inflation of relationships.
where kXTX and kYTY are X’s and Y’s reliability indexes, More directly, Chan (2009) provided quotations from
respectively, qTX TY represents the X–Y true score correla- reviewer comments indicating the assumption that non-
tion, and kXM and kYM represent the effect of the common self-reports are superior to self-reports (p. 325).
method M on X and Y, respectively. Equation 3 shows that One reason self-reports might be considered inferior is
rXY is inflated by a factor of kXMkYM, the common causal because of method effects which decrease construct
effect of method M used to measure both X and Y, and this validity (e.g., influence of response biases on ratings of job
is the source of common method bias. Thus, common characteristics). Self-reports certainly can be subject to
method bias causes observed score correlations to be method effects but literature reviews have indicated that
inflated as compared to their true score counterparts. Note, suspected biasing factors including social desirability
however, that in Eq. 3 rXY is also simultaneously attenu- (Ones et al. 1996) and negative affect and acquiescence
ated by unreliability in both X and Y (i.e., both kXTX and (Spector 2006) do not have strong, consistent effects (e.g.,
kYTY are likely to be [considerably] less than 1.0) as com- Williams and Anderson 1994). In addition, other methods
pared to qTX TY . As such, and perhaps counterintuitively, are also subject to method effects (e.g., supervisor ratings
correlations between variables measured using a common of job characteristics are subject to response biases as well,
method are simultaneously attenuated due to unreliability as are peer ratings of job performance). In fact, if all
and inflated due to common method bias. Thus, the net constructs are measured by the same measurement method
effect could be that observed correlations are (a) higher (e.g., supervisory ratings, peer ratings), the measurement
than qTX TY (if inflation is greater than attenuation), (b) implications are identical to those discussed earlier in
lower than qTX TY (if attenuation is greater than inflation), or connection with Eq. 3 and the case of self-reports.
(c) equal to qTX TY (if inflation and attenuation offset each Another reason for considering self-reports to be inferior
other). is the belief that if two variables are self-reported, the
In a recent study, Lance et al. (in press) re-analyzed data relationship will be inflated (see misconception #1). It
from 18 published multitrait-multimethod (MTMM) might be argued that it is better to measure at least one of
matrices based on a variety of constructs and methods to the variables with an other-report. However, if a relation-
evaluate the possible offsetting effects of unreliability and ship is assessed between a self-reported variable (e.g., job
common method effects. Same-method correlations (rXY in satisfaction) and an other-reported variable (e.g., supervi-
Eq. 3) were compared to trait factor correlations from sor report of job characteristics), the result could be an
confirmatory factor analysis of the MTMM matrices; the attenuated relationship due to unshared method effects (see
trait factor correlations can be regarded as unbiased esti- the appendix and Conway 2002) and/or random error (see
mates of true correlations, qTX TY in Eq. 3. Overall, they Eq. 3 and Lance et al., in press), or even an inflated cor-
found that the average same-method correlation was .342, relation. In the case in which X and Y are measured by
which was actually slightly lower than the mean trait factor different methods Mj and Mj0 , the observed correlation can
correlation of .371. These results suggest that (a) the be represented as:

123
328 J Bus Psychol (2010) 25:325–334

rXY ¼ kXTX kYTY qTX TY þ kXMj kYMj0 qMj Mj0 ð4Þ proportion of variance in performance ratings that is
attributable to traits and that which is attributable to the
where kXMj and kYMj0 represent the effects of the methods Mj methods or raters’’ (p. 559) and that these studies have
and Mj0 on X and Y, respectively, qMj Mj0 reflects the corre- ‘‘documented the ubiquitous phenomenon of method
lation between the different methods, and all other terms effects in performance ratings’’ (p. 568) (see also, Becker
are defined as before. Note that as in Eq. 3, attenuation and Cote 1994; Doty and Glick 1998; Scullen et al. 2000;
incurs to the extent that measures are less than perfectly Scullen et al. 2003, for similar attributions). Combined
reliable (i.e., kXTX and kYTY are less than 1.00), but that with the belief that ‘‘method variance can bias results when
covariance distortion (usually inflation) still incurs to the researchers investigate relations among constructs with the
extent that the methods are positively correlated (i.e., common method… method variance produces a potential
qMj Mj0 [ 0). If the different methods are uncorrelated, then threat to the validity of empirical findings’’ (Bagozzi and
the observed correlation is attenuated due to both unreli- Yi 1990, p. 547), the pervasive and strong rater (source)
ability and uncorrelated method variance as shown in the effects that are routinely found in multisource ratings
appendix. In the event that the methods are negatively (Mount et al. 1998; Scullen et al. 2000, 2003; Hoffman
correlated, then even further attenuation would incur due to et al. 2010) create the impression that these ratings are
the fact that the product kXMj kYMj0 qMj Mj0 in Eq. 4 would be seriously flawed, biased, and inaccurate measures of ratees’
negative, causing rXY to be even less than qTX TY attenuated actual performance. Note that despite misconception #2
simply by measurement error. that other-ratings are superior to self-ratings, beliefs about
We noted earlier that Lance et al. (in press) found even rating sources as methods could lead to criticism of other-
same-method correlations to slightly underestimate true reports as well as self-ratings.
relationships, and this effect was found to be even greater We argue that the mere existence of rater source effects
for correlations between different methods: the mean dif- does not undermine the construct validity of either self-
ferent-method, different-trait correlation was only .248 reports or other reports. In a series of studies, Hoffman,
when compared to the mean trait factor correlation (the Lance and colleagues (Hoffman and Woehr 2009; Lance
estimate of qTX TY ) of .371. In fact, and using results from et al. 2006, 1992) conducted confirmatory factor analysis
Lance et al.’s (in press) Table 2, 61% of the mean differ- (CFA) of multisource performance ratings (MSPRs). As is
ent-method correlation was due to attenuated true-score often the case, they found support for both rating dimen-
correlation (.151/.248), but 39% (.097/.248) was due to sion and rater source factors and found that rater source
effects of measurement methods that were correlated factors accounted for substantially more variance in ratings
qMj Mj0 ¼ :525, on the average. Thus, rather than providing a than did dimension factors. Note that from a traditional
more accurate estimated of true relationships among con- perspective, this pattern of evidence would be interpreted
structs, relationships estimated using different methods as representing the presence of substantial rater (qua
tend to be more attenuated and less accurate as compared method) bias. However, Hoffman, Lance and colleagues
to same-method correlations. This is true despite the fact also correlated rater source factors with additional job-
that different-method correlations reflect some inflation related variables external to the ‘‘core’’ CFA, including
due effects of correlated methods. In fact, had the different personality, aptitude, experience, and objective perfor-
methods been uncorrelated, the average different-method mance measures to test whether source factors represented
correlation would have been rXY = .151, rather severely mere performance-irrelevant rater (i.e., method) bias fac-
attenuated as compared to its true-score counterpart tors or whether, as an ecological perspective on ratings (see
(qTX TY ¼ :371). Lance et al. 2006) and MSPR practitioners (e.g., Tornow
1993) suggest, rater source factors represent alternative,
Misconception #3: Rating Sources (e.g., Self, Other) and complementary valid perspectives on ratee perfor-
Constitute Measurement Methods mance. That rater source factors evidenced significant
relationships with other performance-related variables
Underlying the preference for other-reports over self- supported the latter position. For example, Lance et al.
reports (see ‘‘Misconception #2’’) is an assumption that (1992) showed that a self-rating factor correlated with
different rating sources, for example, supervisor, peer, mechanical aptitude and job experience, and did so more
subordinate, and self-ratings in multisource rating systems, strongly than did supervisor or peer ratings. These results
represent different methods for measuring ratee perfor- (a) support the construct validity of self-ratings, and (b)
mance. For example, Mount et al. (1998) wrote that more generally show that different rater sources, including
‘‘[s]tudies that have examined performance rating data self-ratings, represent valid performance-related variance
using multitrait-multimethod matrices (MTMM) or multi- and not mere measurement method bias (see reviews by
trait-multirater (MTMR) matrices usually focus on the Lance et al. 2008, 2009).

123
J Bus Psychol (2010) 25:325–334 329

What Reviewers Should Reasonably Expect perceived by the incumbent), making self-reports the theo-
from Authors Regarding Common Method Bias retically most relevant measurement method. First, they
used the term ‘‘perceived job characteristics’’ throughout the
To a certain extent it is easier to say what we think article. Second, their rationale was tied to perceptions of job
reviewers and other gatekeepers should not do (i.e., base characteristics rather than objective job characteristics. For
criticisms on the misconceptions described earlier) than it example, they argued that those who dispositionally expe-
is to prescribe proper reviewer behavior. It is not possible rience positive affect may respond more favorably to situ-
to ensure that method effects do not influence results ations (i.e., perceive a given job more positively) than those
(indeed, it is widely accepted that the act of measurement who tend to experience negative affect.
can and does change the phenomenon being studied; Hei-
senberg 1927), but it is reasonable for reviewers to expect Construct Validity Evidence
authors to take certain steps to reduce the likelihood of
common method bias. We recommend four things One way to rule out substantial method effects is to dem-
reviewers can look for, but we emphasize that it is not our onstrate construct validity of the measures used. Authors
intent to define hurdles for authors that reviewers and other should be able to make an argument that the measures they
gatekeepers should feel bound to enforce during the review chose have construct validity, and it is reasonable for
process. Overreliance on rules of thumb without careful reviewers to ask for it if it is not provided. On the other
consideration of the origin and rationale for the rules and hand, we do not mean to imply that authors should always
their application is very problematic, and is instrumental in be required to use the most widely cited and most thor-
creating and perpetuating what Vandenberg (2006) and oughly researched measures as we believe that this could
Lance and Vandenberg (2009) have referred to as ‘‘statis- retard creative research toward refinement of theoretical
tical and methodological myths and urban legends.’’ With constructs and their operationalizations. We also do not
this in mind, we believe it is reasonable for reviewers to want to prescribe a specific set of criteria that authors
expect the following: should meet because (a) establishing construct validity
depends on a constellation of types of evidence (Putka and
An Argument for Why Self-Reports are Appropriate Sackett 2010), (b) it is not reasonable for researchers to
provide extensive evidence in every case, especially when
Self-reports are clearly appropriate for job satisfaction and the area of research is relatively new and innovative, and
many other private events (Chan 2009; Skinner 1957), but (c) the availability and most appropriate type of evidence
for other constructs such as job characteristics or job per- may vary case by case. Rather, we list several types of
formance, other types of measures might be appropriate or evidence likely to be useful for establishing construct
even superior. If reviewers have a specific concern about validity; whether the evidence in a given case is adequate is
the use of self-reports (e.g., a concern about social desir- a matter of judgment.
ability in self-reports on sensitive topics) they should The kinds of evidence likely to be useful include: (a)
expect that authors/researchers be able to provide solid appropriate reliability evidence (Feldt and Brennan 1989),
rationale for their choice. (b) factor structure, (c) establishment of a nomological
An excellent example of a rationale is provided by network including relationships with theoretically relevant
Shalley et al. (2009) who studied self-reported creative variables (or lack of relationships where this is expected
performance as predicted by growth need strength and theoretically), (d) mean differences between relevant
work context variables. Their rationale for using self- groups, (e) convergent and discriminant validity based on a
reports of creativity included an argument for why self- multitrait-multimethod matrix, or (f) other evidence
reports are particularly appropriate, stating ‘‘it has been depending on the situation (Messick 1989). Oftentimes,
argued that employees are best suited to self-report crea- this type of evidence is available in sources that describe
tivity because they are the ones who are aware of the subtle the initial development and evaluation of the particular
things they do in their jobs that make them creative’’ measure (see, e.g., Dooley et al. 2007; Mallard and Lance
(p. 495). They also cited evidence from another study that 1998; Netemeyer et al. 1996). In these cases, reviewers
self-ratings of creativity correlated substantially (r = .62) could reasonably expect authors to acknowledge the scale
with supervisor ratings. development citation and perhaps provide supplementary
Another best-practice example is provided by Judge et al. corroborative evidence of the scale’s internal consistency
(2000), who studied job characteristics as a mediator of the and factor structure.
relationship between core self evaluations and job satisfac- For cases in which extensive construct validity evi-
tion. In describing their theoretical model Judge et al. dence has not been presented elsewhere, reviewers might
explicitly focused on perceived job characteristics (i.e., as reasonably ask for additional evidence. Examples of

123
330 J Bus Psychol (2010) 25:325–334

reasonable construct validity evidence include Shalley emphasize that ruling out conceptual overlap should not be
et al. (2009) and MacKenzie et al. (1998). Shalley et al. a new hurdle that all authors are expected to clear. Rather,
justified self-ratings of creativity by noting that self-ratings it makes sense to raise the issue if a reviewer has a specific
had been shown in other research to correlate substantially reason to believe overlap exists.
(r = .62) with supervisor ratings, in addition to providing
evidence of internal consistency reliability. MacKenzie Evidence that Authors Proactively Considered
et al. (1998) used instruments with long track records such Common Method Bias
as the Minnesota Satisfaction Questionnaire, and cited
multiple sources providing construct validity evidence. It is reasonable for reviewers to look for evidence that
authors considered common method bias in the design of
Lack of Overlap in Items for Different Constructs their study. This evidence can take many forms, and we do
not recommend that authors be required to use any par-
Conceptual overlap in items used to measure different ticular one—merely that the judicious use of design tech-
constructs can bias relationships (Brannick et al., in press). niques can demonstrate an a priori consideration of method
This is something reviewers should watch out for. It can be effects. Podsakoff et al. (2003) provided examples of
difficult because usually authors do not, understandably, design techniques, including ‘‘temporal, proximal, psy-
provide all items for their measures. But, we believe it is chological, or methodological separation of measurement,’’
good to be alert for this kind of thing, and reviewers can ‘‘protecting respondent anonymity and reducing evaluation
ask authors about potential overlap if the constructs are apprehension,’’ ‘‘counterbalancing question order,’’ and
similar. Possible remedies could include rewording or ‘‘improving scale items’’ (pp. 887–888).
dropping items that do overlap. We believe any of these can demonstrate a priori con-
One research area for which this issue has been inves- sideration of the possibility of common method bias,
tigated is the relationship between organizational com- though it might be demonstrated in other ways. The
mitment and turnover intentions. Bozeman and Perrewé appropriate technique(s) will depend on the situation. One
(2001) had expert judges rate items from the organizational good example of a priori consideration comes from Wes-
commitment questionnaire (OCQ; Mowday et al. 1979) taby and Braithwaite (2003), who studied reemployment
and concluded that several OCQ items were retention- self-efficacy. Westaby and Braithwaite counterbalanced
related (e.g., ‘‘It would take very little change in my sets of items, and clearly explained the procedure and
present circumstances to cause me to leave this organiza- rationale. Specifically, their purposes were to ‘‘(a) mitigate
tion’’). They used confirmatory factor analysis to show that order effects…, (b) reduce the negative effect of item order
these items loaded significantly on a turnover-cognition on theoretical testing…, and (c) reduce the potential for
factor (along with a set of turnover cognition items), and response sets’’ (p. 425). They used four different item
that the OCQ-turnover cognition correlation decreased orders with random assignment of respondents to orders.
when the overlapping OCQ items were excluded. Another example might be measuring the constructs
While Bozeman and Perrewé’s (2001) explicit purpose believed to contribute to common method bias, and
was to study item overlap, Chan (2001) acknowledged and incorporating them into a structural equation model, as
dealt with the issue in a substantive study of the organi- Williams and Anderson (1994) did. However, as we note
zational commitment (OC)–turnover intentions (QU) rela- later, while this approach is theoretically sound it has not
tionship. Chan stated that ‘‘OC was measured using a six- been demonstrated to reduce common method bias.
item scale adapted from Mowday et al. (1979). The six
items made no reference to intent to quit (QU) to avoid
artifactual inflation of OC–QU correlation due to item A Note on Post Hoc Statistical Control of Common
overlap’’ (p. 81). Method Bias
We suspect there are other constructs for which items
are likely to overlap. One example involves measures of One recommendation that we did not include in our list was
contextual performance and the personality trait consci- post hoc statistical control of common method bias. A
entiousness, two constructs that have been theoretically number of techniques have been proposed and thoroughly
linked (Motowidlo et al. 1997). Van Scotter and Mot- reviewed by Podsakoff et al. (2003). One of these is Lindell
owidlo’s (1996) contextual performance measure (specifi- and Whitney’s (2001) partial correlation approach in which
cally the job dedication subscale) included the item ‘‘pay the minimum correlation between some ‘‘marker variable’’
close attention to important details’’ and a measure of and a study’s focal variables is subtracted from the corre-
conscientiousness from the International Personality Item lations among the focal variables to adjust for common
Pool (2010) is ‘‘Pay attention to details.’’ We want to method bias. According to Richardson et al. (2009) this

123
J Bus Psychol (2010) 25:325–334 331

approach has been frequently adopted (48 times as of June and for different-method correlations is:
2008), but usually to demonstrate that method bias was not rXY  kXMj kYMj0 qMj Mj0
a problem in the first place. ^ TX TY ¼
q ð6Þ
kXTX kYTY
A second approach involves loading all study measures
onto an unmeasured ‘‘method’’ factor and then analyzing where all terms are defined as in Eqs. 3 and 4 above. As
relationships among the residualized variables. According such, these corrections simultaneously adjust observed
to Richardson et al. (2009), at least 49 studies through June correlations (rXY) for attenuation due to unreliability
2008 have used the unmeasured method factor approach. (kXTX kYTY in the denominator) and for inflation due to
We see this approach as logically indefensible as it may method effects (kXM kYM or kXMj kYMj0 qMj Mj0 in the numera-
easily remove trait variance when multiple traits have a tor). Lance et al. (in press) also present a multivariate
common cause (e.g., generalized psychological climate; extension of these corrections and provide discussion of
James and James 1989). approaches to obtaining estimates of the quantities needed
Both the partial correlation and the unmeasured method to effect the corrections. While this approach appears to be
techniques were evaluated in a simulation by Richardson psychometrically sound there has been no empirical work
et al. (2009) and both produced disappointing results. In at all evaluating its effectiveness.
fact, they produced less accurate estimates of correlations Thus, post hoc statistical control of common method
than did no correction at all when (a) method variance was bias is in an unfortunate state-of-the-practice: some well-
present and (b) true correlations were greater than zero— known and easily applied approaches appear to be inef-
conditions that are typical of organizational research. We fective, use of the MTMM matrix is routinely resource
therefore discourage reviewers from suggesting that intensive, even if multiple measurement methods can be
authors use either of these commonly used remedies for identified for focal constructs, and the efficacy of Lance
common method bias. et al.’s (in press) statistical control procedure is as yet
A third approach is Williams and Anderson’s (1994) untested. Consequently, we cannot recommend any of
technique in which all focal latent variables’ manifest these approaches.
indicators are loaded onto one or more substantive method
latent variables (e.g., acquiescence, social desirability,
A Counter-Argument
positive/negative affectivity) on which also load the
method factor’s manifest indicators. One limitation of this
Two readers of an earlier version of our manuscript inde-
technique is that it presumes that a researcher knows the
pendently suggested that, while our argument about atten-
source of method variance. The approach appears to have
uation in relationships due to use of different methods may
been seldom applied and to have little effect on estimates
be valid, use of different methods is still preferable to the use
among the focal constructs (Richardson et al. 2009).
of the same method for different constructs. The argument is
Two other approaches appear promising, though each
that different—versus same-method studies may lead to
has important limitations. Use of a multitrait-multimethod
different types of errors—using different methods will tend
(MTMM) matrix to estimate relationships can be very
to underestimate relationships while using the same method
effective for controlling common method bias (if methods
may tend to overestimate relationships (however; see our
are chosen carefully; see our ‘‘Misconception #3’’). But the
‘‘Misconception #1’’). Thus, according to this argument, it is
substantial demands of carrying out a MTMM study make
more impressive to find a significant relationship with dif-
it difficult or impossible in many research situations.
ferent methods because it is harder to do so. In this regard,
Although there are thousands of citations to Campbell and
one reader made an analogy to Type-I errors being more
Fiske’s (1959) original article on the MTMM methodology
often associated with same-method correlations (same-
and hundreds of applications of the MTMM matrix (Lance
method correlations may overestimate relationships) versus
et al., in press) we know of only one application in which
Type-II errors being more often associated with different-
researchers tested directional relationships (i.e., a structural
method correlations (different-method correlations may
model) among self-reported constructs while controlling
underestimate relationships).
for method effects in the MTMM matrix (Arora 1982).
We acknowledge this analogy and we agree that in
A second promising approach is Lance et al.’s (in press)
general it is more difficult to detect a significant relation-
simultaneous correction for common method bias and
ship between constructs measured by different methods
attenuation. The correction for same-method correlations is:
(i.e., higher probability of a Type-II error). However, we do
rXY  kXM kYM
^ TX TY ¼
q ð5Þ not agree that this is a virtue; problems with protecting
kXTX kYTY against Type-I error at the expense of a greater Type-II

123
332 J Bus Psychol (2010) 25:325–334

error rate have been discussed, for example, by Hunter will actually attenuate correlations as compared to the
(1997) and Cohen (1994). It is desirable to have high power no-method-variance situation. First consider a situation
in research (i.e., to have a low Type-II error rate), attenu- with no method variance in measures of two traits (TX and
ated relationships reduce power by increasing the likeli- TY) which are measured by two different, uncorrelated
hood of Type-II errors, and Lance et al.’s (in press) review methods (Mj and Mj0 ). The equation for the correlation
indicates that different-method correlations are routinely coefficient, expressed in terms of variances and covariance,
attenuated as compared to their true-score counterparts. when no method variance is present, is:
We do acknowledge the potential in some cases for COVðTX ; TY Þ
inflation due to common method bias (though Lance et al.’s rxy ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffipffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
VARðTX Þ þ VARðEX Þ VARðTY Þ þ VARðEY Þ
[in press] review indicates that same-method correlations
are not routinely upwardly biased), and we are certainly not ðA1Þ
opposed to the use of different methods to estimate rela- Now consider what happens when each measure contains
tionships. But, we believe that if reviewers wish to criticize a method variance that is unshared with the other measure.
study for using a same-method design, there should be a The variance terms for each measure’s method effect,
rationale for why the use of the same method would be VAR(Mj) and VAR(Mj0 ), are added to the denominator:
problematic. Reviewers (as well as authors) should take a Because additional method variance appears in the
thoughtful approach to the issue of measurement methods; denominator when there are method effects (Eq. 8), but not
each research situation’s characteristics (e.g., potential for when there are no method effects (Eq. 7), Eq. 8 denomi-
social desirability, availability of valid measures, etc.) nator will be larger while the numerators are the same. The
should be considered and a determination made based on correlation value in Eq. 8 will therefore be attenuated.

COVðTX ; TY Þ
rxy ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffipffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðA2Þ
VARðTX Þ þ VARðMj Þ þ VARðEX Þ VARðTY Þ þ VARðMj0 Þ þ VARðEY Þ

those characteristics, rather than on a general notion that


same-method correlations are necessarily problematic.

References
Summary
Arora, R. (1982). Validation of an S-O-R model for situation,
We believe that journal reviewers and editors, and other enduring and response components of involvement. Journal of
Marketing Research, 19, 505–516.
gatekeepers, such as dissertation or thesis committee Bagozzi, R. P., & Yi, Y. (1990). Assessing method variance in
members, have to some extent perpetuated misconceptions multitrait-multimethod matrices: The case of self-reported affect
about common method bias in self-report measures, for and perceptions at work. Journal of Applied Psychology, 75,
example, that relationships between self-reported variables 547–560.
Becker, T. E., & Cote, J. A. (1994). Additive and multiplicative
are necessarily and routinely upwardly biased. We argued method effects in applied psychological research: An empirical
against these misconceptions and recommended what assessment of three models. Journal of Management, 20, 625–
gatekeepers should reasonably expect regarding common 641.
method bias. Reasonable expectations include (a) an argu- Bozeman, D., & Perrewé, P. (2001). The effect of item content
overlap on Organizational Commitment Questionnaire–turnover
ment for why self-reports are appropriate, (b) a case for the cognitions relationships. Journal of Applied Psychology, 86,
construct validity of the measures, (c) lack of overlap in 161–173.
items for different constructs, and (d) proactive measures on Brannick, M. T., Chan, D., Conway, J. M., Lance, C. E., & Spector,
the part of authors to minimize threats of common method D. (in press). What is method variance and how can we cope
with it? A panel discussion. Organizational Research Methods.
bias. No post hoc statistical correction procedure can be Campbell, J. (1982). Editorial: Some remarks from the outgoing
recommended until additional research evaluates the rela- editor. Journal of Applied Psychology, 67, 691–700.
tive effectiveness of those that have been proposed. Campbell, D. T., & Fiske, D. W. (1959). Convergent and discriminant
validation by the multitrait-multimethod matrix. Psychological
Bulletin, 56, 81–106.
Appendix: Attenuation of Correlations by Unshared Chan, D. (2001). Method effects of positive affectivity, negative
Method Variance affectivity, and impression management in self-reports of work
attitudes. Human Performance, 14, 77–96.
Chan, D. (2009). So why ask me? Are self-report data really that bad?
Method variance is usually assumed to inflate correlations, In C. E. Lance & R. J. Vandenberg (Eds.), Statistical and
but in a fairly common research situation, method variance metholodogical myths and urban legends: Doctrine, verity and

123
J Bus Psychol (2010) 25:325–334 333

fable in the organizational and social sciences (pp. 311–338). Lance, C. E., Teachout, M. S., & Donnelly, T. M. (1992).
New York: Routledge. Specification of the criterion construct space: An application of
Cohen, J. (1994). The earth is round (p \ .05). American Psychol- hierarchical confirmatory factor analysis. Journal of Applied
ogist, 49, 997–1003. Psychology, 77, 437–452.
Conway, J. M. (2002). Method variance and method bias in I/O Lance, C. E., & Vandenberg, R. J. (2009). Introduction. In C. E.
psychology. In S. G. Rogelberg (Ed.), Handbook of research Lance & R. J. Vandenberg (Eds.), Statistical and metholodogical
methods in industrial and organizational psychology (pp. 344– myths and urban legends: Doctrine, verity and fable in the
365). Oxford: Blackwell Publishers. organizational and social sciences (pp. 1–4). New York:
Cote, J. A., & Buckley, M. R. (1987). Estimating trait, method and Routledge.
error variance: Generalizing across 70 construct validation Lindell, M. K., & Whitney, D. J. (2001). Accounting for common
studies. Journal of Marketing Research, 26, 315–318. method variance in cross-sectional research designs. Journal of
Crampton, S., & Wagner, J. (1994). Percept-percept inflation in Applied Psychology, 86, 114–121.
microorganizational research: An investigation of prevalence Lord, F. M., & Novick, M. R. (1968). Statistical theories of mental
and effect. Journal of Applied Psychology, 79, 67–76. test scores. Reading, MA: Addison-Wesley Publishing
Dooley, W. K., Shaffer, D. R., Lance, C. E., & Williamson, G. M. Company.
(2007). Informal care can be better than adequate: Development MacKenzie, S. B., Podsakoff, P. M., & Ahearne, M. (1998). Some
and evaluation of the exemplary care scale. Rehabilitation possible antecedents and consequences of in-role and extra-role
Psychology, 52, 359–369. salesperson performance. Journal of Marketing, 62, 87–98.
Doty, D. H., & Glick, W. H. (1998). Common method bias: Does Mallard, A. G. C., & Lance, C. E. (1998). Development and
common methods variance really bias results? Organizational evaluation of a parent–employee interrole conflict scale. Social
Research Methods, 1, 374–406. Indicators Research, 45, 343–370.
Feldt, L. S., & Brennan, R. L. (1989). Reliability. In R. L. Linn (Ed.), Messick, S. (1989). Validity. In R. L. Linn (Ed.), Educational
Educational measurement (3rd ed., pp. 105–146). Washington, measurement (pp. 13–103). Washington, DC: American Council
DC: American Council on Education and National Council on on Education and National Council on Measurement in
Measurement in Education. Education.
Heisenberg, W. (1927). Über den anschaulichen Inhalt der quanten- Motowidlo, S., Borman, W., & Schmit, M. (1997). A theory of
theoretischen Kinematik und Mechanik. [About the graphic individual differences in task and contextual performance.
content of quantum theoretic kinematics and mechanics]. Human Performance, 10, 71–83.
Zeitschrift für Physik, 43, 172–198. Mount, M. K., Judge, T. A., Scullen, S. E., Sytsma, M. R., & Hezlett,
Hoffman, B. J., Lance, C. E., Bynum, B. H., & Gentry, W. A. (2010). S. A. (1998). Trait, rater, and level effects in 360-degree
Rater source effects are alive and well after all. Personnel performance ratings. Personnel Psychology, 51, 557–576.
Psychology, 63, 119–151. Mowday, R. T., Steers, R. M., & Porter, L. W. (1979). The
Hoffman, B. J., & Woehr, D. J. (2009). Disentangling the meaning of measurement of organizational commitment. Journal of Voca-
multisource performance rating source and dimension factors. tional Behavior, 14, 224–247.
Personnel Psychology, 62, 735–765. Netemeyer, R. G., Boles, J. S., & McMurrian, R. (1996).
Hunter, J. (1997). Needed: A ban on the significance test. Psycho- Development and validation of work-family conflict and
logical Science, 8, 3–7. family-work conflict scales. Journal of Applied Psychology,
International Personality Item Pool. (2010). International Personality 81, 400–410.
Item Pool: A scientific collaboratory for the development of Ones, D., Viswesvaran, C., & Reiss, A. (1996). Role of social
advanced measures of personality traits and other individual desirability in personality testing for personnel selection: The red
differences (http://ipip.ori.org/). http://ipip.ori.org/newNEOKey. herring. Journal of Applied Psychology, 81, 660–679.
htm#Conscientiousness. February 5, 2010. Organ, D., & Ryan, K. (1995). A meta-analytic review of attitudinal
James, L. A., & James, L. R. (1989). Integrating work environment and dispositional predictors of organizational citizenship behav-
perceptions: Explorations into the measurement of meaning. ior. Personnel Psychology, 48, 775–802.
Journal of Applied Psychology, 74, 739–751. Podsakoff, P., MacKenzie, S., Lee, J., & Podsakoff, N. (2003).
Judge, T., Bono, J., & Locke, E. (2000). Personality and job Common method biases in behavioral research: A critical review
satisfaction: The mediating role of job characteristics. Journal of of the literature and recommended remedies. Journal of Applied
Applied Psychology, 85, 237–249. Psychology, 88, 879–903.
Lance, C. E., Baranik, L. E., Lau, A. R., & Scharlau, E. A. (2009). If Podsakoff, P., & Todor, W. (1985). Relationships between leader
it ain’t trait it must be method: (Mis)application of the multitrait- reward and punishment behavior and group processes and
multimethod design in organizational research. In C. E. Lance & productivity. Journal of Management, 11, 55–73.
R. J. Vandenberg (Eds.), Statistical and metholodogical myths Putka, D. M., & Sackett, P. R. (2010). Reliability and validity. In J. L.
and urban legends: Doctrine, verity and fable in the organiza- Farr & N. T. Tippins (Eds.), Handbook of employee selection
tional and social sciences (pp. 339–362). New York: Routledge. (pp. 9–49). New York: Routledge.
Lance, C. E., Baxter, D., & Mahan, R. P. (2006). Multi-source Richardson, H., Simmering, M., & Sturman, M. (2009). A tale of
performance measurement: A reconceptualization. In W. Ben- three perspectives: Examining post hoc statistical techniques for
nett, C. E. Lance, & D. J. Woehr (Eds.), Performance detection and correction of common method variance. Organi-
measurement: Current perspectives and future challenges (pp. zational Research Methods, 12, 762–800.
49–76). Mahwah, NJ: Erlbaum. Scullen, S. E., Mount, M. K., & Goff, M. (2000). Understanding the
Lance, C. E., Dawson, B., Birklebach, D., & Hoffman, B. J. (in press). latent structure of job performance ratings. Journal of Applied
Method effects, measurement error, and substantive conclusions. Psychology, 85, 956–970.
Organizational Research Methods. Scullen, S. E., Mount, M. K., & Judge, T. A. (2003). Evidence of the
Lance, C. E., Hoffman, B. J., Baranik, L. E., & Gentry, W. A. (2008). construct validity of developmental ratings of managerial
Rater source factors represent important subcomponents of the performance. Journal of Applied Psychology, 88, 50–66.
criterion construct space, not rater bias. Human Resource Shalley, C., Gilson, L., & Blum, T. (2009). Interactive effects of
Management Review, 18, 223–232. growth need strength, work context, and job complexity on self-

123
334 J Bus Psychol (2010) 25:325–334

reported creative performance. Academy of Management Jour- Westaby, J., & Braithwaite, K. (2003). Specific factors underlying
nal, 52, 489–505. reemployment self-efficacy: Comparing control belief and
Skinner, B. F. (1957). Verbal behavior. New York: Appleton- motivational reason methods for the recently unemployed.
Century-Crofts. Journal of Applied Behavioral Science, 39, 415–437.
Spector, P. (2006). Method variance in organizational research: Truth Widaman, K. (1985). Hierarchically nested covariance structure
or urban legend? Organizational Research Methods, 9, 221–232. models for multitrait-multimethod data. Applied Psychological
Tornow, W. W. (1993). Perceptions or reality: Is multi-perspective Measurement, 9, 1–26.
measurement a means or an end? Human Resource Management, Williams, L. J., & Anderson, S. E. (1994). An alternative approach to
32, 221–229. method effects by using latent-variable models: Applications in
Van Scotter, J. R., & Motowidlo, S. J. (1996). Interpersonal organizational behavior research. Journal of Applied Psychol-
facilitation and job dedication as separate facets of contextual ogy, 79, 323–331.
performance. Journal of Applied Psychology, 81, 525–531. Williams, L. J., Cote, J. A., & Buckley, M. R. (1989). Lack of method
Vandenberg, R. J. (2006). Statistical and methodological myths and variance in self-reported affect and perceptions at work: Reality
urban legends: Where, pray tell, did they get this idea? or artifact? Journal of Applied Psychology, 74, 462–468.
Organizational Research Methods, 9, 194–201.

123
http://www.springer.com/journal/10869

View publication stats

Você também pode gostar