TU-L0022 - Statistical Research Methods D, Lecture, 2.11.2021-6.4.2022
This course space end date is set to 06.04.2022 Search Courses: TU-L0022
Systematic measurement error and common method variance (14:49)
Description to be added.
This video tells what common method variance is about.
Click to view transcript
Random
noise is not the only kind of measurement error that researchers need
to worry about. Particularly if you're doing Survey Research then you
need to be aware of the issue of common method variance. Whether or not
you think that common method variance could influence your results some
of your readers and some of your reviewers will think that that could be
the case and therefore you have to address this issue head-on if you do
a cross-sectional survey. Let's take a look at what common method
variance is about. Here's an example. So we have three questions that
are supposed to measure innovativeness and three questions that are
supposed to measure performance of a company or success of a company. Let's
take a look at our example. So how would we improve this possibly
common method variance contaminated survey form? We could of course do a
second survey where we ask the performance or success implications
asked half a year later, one year later, but even better if we study
companies we don't actually have to measure this success with a survey.
We can rely on accounting data. So we wait two years then we get the
actual accounting data - what kind of growth or kind of innovative what
kind of profitability the company actually reported and then we compare
whether those companies that are innovative grew more or were more
profitable than less innovative companies. One good strategy for
implementing this kind of study is that oftentimes when you write a
paper - you first write it to a conference and then you write the better
version that you try to publish in a journal - is that you can collect
the success data or performance data using a survey first and then you
write the paper using the survey data as the dependent variable for the
conference. Then you get some feedback and maybe one year later you have
- we see the conference gotten feedback - then you start to work toward
the journal paper. When the journal paper is done then by that time you
will have the actual accounting measures available because it typically
takes a year or two from the idea to a publishable journal paper if you
do the conference presentation in-between. So you can do first the
conference with survey data then switch to a better dependent variable
by using actual accounting data and this also helps you to avoid the
concern that you're just trying to republish the same study that you
presented in a conference in a journal because when you have a different
dependent variable then no one can argue that it's the same study
anymore.
So
we have innovation questions i1 i2 and i3 - we have success questions
s1 s2 and s3. We follow the common practice of taking a sum of these i
indicators and some of these success indicators. We find that the
correlation of the sum of the innovation indicators and some of the
success indicators is 0.3 and we assume that these innovation indicators
measure innovativeness these success indicators measure success and
there could be some other variance components or random noise and some
either specific variation that we don't really care about in the data.
So we find correlation 0.3. We know that random measurement error
attenuates correlations so we claim that the real correlation could
actually be as high as 0.4 and then we make grand claims about
innovation being one of the key drivers of success. We claim innovation
and success must be associated. So what's the - what kind of problems do
we have? Do we have any alternative explanations for the correlation?
It's also possible that these indicators don't measure only
innovativeness and only success but they measure also whether the person
thinks positively about the company. So we have this systematic
measurement error S here and it influences all these indicators. So a
skeptic of our study would say that we have not found out that
innovation and success are associated instead we have just found out
that when we present positive statements of a company to a person then
some people will respond systematically higher than others. So that is
instead of being driven by the constructs these are driven by general
sentiment of the person and they don't really - the correlation is not
really a reflection of any theoretical relationship.
Let's take
another example. This is from a paper published in Information Systems
Research. And we have a question scales about information quality. We
have a scale about accuracy and we have a scale about completeness of
information. And these are about government information systems. Now the
question is that measures correlate accuracy 1 and completeness 1
correlate. So what can we say? We can say that they measure two
different constructs and the constructs are correlated. A skeptic would
say that no these indicators don't measure two different constructs.
Instead they measure hostility toward government. So particularly if you
do this in the United States - where I think this research was done -
there are people who really think that government shouldn't be doing any
services for people and they are openly hostile about government. So if
you're hostile toward government you're gonna rate all these indicators
to small numbers and if you like government services you'll rate them
to high numbers regardless of the accuracy of data and completeness of
the data. These or they could just measure how much the informant wants
to answer ones versus fives when asked to agree on an item. So that's
also possible.
This issue is called common method variance. So
the idea of common method variance is that the correlation between two
indicators is not driven by the correlation between the constructs but
instead it's an outcome of the measurement process. So some people like
to measure - answer the center of the scale some people like to answer
the ends and that causes a small correlation. Whether the correlation
between indicators is entirely or partly influenced by method variance
is one issue. But the thing is that if a reviewer challenges you that
you have this problem you have to be able to demonstrate that you don't
or it's also possible that you have it. In which case you have to
understand how to avoid it. And to avoid the problem and argue why we
wouldn't have a common method variance plot problem if we really believe
so we have to understand the different sources of method variance. So
why does a method or how a measurement method can induce variation into
our data.
There's a good paper or actually the series papers
written by Philip Podsakoff and co-authors and they have this 2003 paper
which is the most cited one. They have this big table where they list
all kinds of different reasons why survey indicators could be
correlated. So it is possible that indicators are correlated because you
asked them from the same person. So if you have have the same person
responding to your innovation questions and success questions then that
can cause a correlation between those indicators because of people's
tendency to respond survey questions. It's also possible that there are
social desirability bias. So some indicators some items are - it's very
difficult to agree on. For example if you ask somebody whether they have
committed a crime or not. It's very difficult to say that you have
committed a crime. And even if you ask the same person whether they have
driven over the speed limit then they would agree more easily because
there's less social desirability to agree on that indicator. Then there
are also item context effects. So if you ask the question first that
makes people angry then that will influence all subsequent indicator
responses. Or if you have a question of whether the company is
innovative or not and then you have indicators after that that measure
specific aspects or specific consequences of innovativeness then the
general question will prime the person to answer positively or
negatively to the remaining questions depending on how they answer the
first question. And then there's the measurement context effects. Like
some people want to answer in a specific way when using paper and pen.
Some people answer in a specific way in online forms and that can cause
various. So there are many many different ways why your survey
indicators can become correlated that is not due to the theoretical
constructs. Is this a big problem then? Well there's quite a lot of
disagreement.
The big problem - whether you think it's a big
problem or not - is not as relevant as the fact that if you do survey
based studies some of your readers and reviewers will think that it's a
big problem and they have studies that they can cite to demonstrate that
this is actually a big thing. For example in this paper by Podsakoff in
2012 they indicated - they reviewed studies that assess the prevalence
of method variance. So how much of the variation in our indicators are
due to the measurement method and how much due to the actual
construction trade being measured and they found out that in the method
that they applied revealed that about one-fifth or one-fourth of the
variance in the data is due to the method. About half or bit less is due
to the trait. So the method variance is about half of the variation of
the actual trait. So if your reliable variation then this implies that
of that reliable variation about sixty-six percent is the actual thing
that you want to measure about 33 is a method variance and then
remaining is the randomness or unreliability. You can of course
challenge this kind of studies based on questioning their methods and so
on but the the point is that this is a potential problem and if a
reviewer says that you have a problem it is difficult to address it. And
oftentimes they do.
We have to understand beyond the problem
itself you have to understand that it has -if it exists - it has serious
consequences. In random measurement - the random noise here - we are
discussed before that if we have a real correlation of 0.5 then
perfectly perfect measurement gives us a very close estimate of the true
correlation if our sample size is large enough. Here is 300 which
definitely large enough. If we have random noise in the data then the
correlations will be underestimated or attenuated. Here the reliability
is 50% and this is attenuated by about minus 40%. So instead of 0.5 we
have 0.29. Systematic measurement error here- we have a systematic error
here - is more problematic because it inflates the correlation
estimate. So in this case x and y both measure - the latent x and y that
were interested in - and there is a systematic measurement error source
that is equally strong as x and y. So observed X is half of the
systematic error and half of the latent X. So it's a 50% systematic
error 50% construct variance. And that will cause serious over
estimation of the relationships. So in this case the relationship is
overestimated by more than 50%. So the real correlation is 0.5 and we're
estimated at 0.77. The common method variance problem is a big deal
because - not only because it inflates existing correlations - it will
also indicate that correlations exist when in reality they do not. So in
this case even if x and y would be completely uncorrelated the
estimated correlation would be in around zero point five ballpark. So
that's your concluding a strong effect when none exists. And this is the
reason why we can deal with unreliability. So unreliability we just
know that the effects are on average a bit smaller than what they should
be. But here with systematic measurement error the problem is that we
can find substantially large effects when none exists and therefore this
is a big problem. It's certainly an important issue that some journals
are actively discouraging cross-sectional surveys. So of course when you
have a survey study where you are using the same scale format at the
same occasions in the same person you ask three questions. Then they
could be correlated and that is one of the reasons why journals are
recommended against cross-sectional studies. So the recommendation is
that instead of doing a cross-sectional study do a study where you
measure on the independent variables with a survey and then measure the
dependent variable using some other means. If nothing else then at least
use a second survey. This of course helps also if you use two surveys.
It helps you to establish the causal order by measuring the effect after
the cause. So that allows you to take into account or argue the second
condition for causality.