SPSS Guide and Normal Values
SPSS Guide and Normal Values
discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/262151892
Introduction to SPSS
CITATIONS READS
2 10,520
1 author:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Research project: Goods & services tax noncompliance model View project
All content following this page was uploaded by Saiyidi Mat Roni on 05 June 2014.
The user has requested enhancement of the downloaded file. All in-text references underlined in blue are added to the original document
and are linked to publications on ResearchGate, letting you access and read them immediately.
Saiyidi MAT RONI
2014
Introduction to SPSS
Files used in the exercises are datasets tailored specifically for the
illustration purposes.
Saiyidi
RONI
SOAR Centre DN: cn=Saiyidi MAT RONI,
Graduate Research School o=Edith Cowan University, ou,
MAT RONI
email=m.matroni@ecu.edu.au,
Edith Cowan University c=AU
Joondalup, Australia. Date: 2014.06.05 16:29:36
+08'00'
Table of Contents
MVA Choose MI or EM
Check transform if
necessary
normality
Cronbach's alpha
Reliability AVE
In the early days of the computer, there was an infamous term used to describe
characteristics of the number crunching machine GIGO! Its an acronym for garbage in garbage
out, which means that no matter how smart the computer is, if the input is wrong, a wrong output
will eventually suffice. Thats a fair description for most computer processes.
SmartPLS, SPSS, AMOS and other statistical software will process your data. But consider
this. If you have a kilo of garbage and later put it on a bathroom scales. Youll get a kilo of garbage
sitting on a really sophisticated measuring instrument.
However, if you pre-sort and pre-sift the garbage into appropriate categories, and then you
measure it accordingly, youll get something more valuable, more meaningful and most importantly
more useful and VALID. For example, if you sort and sift that 1 kilogram of garbage and you get 600
grams of papers and 400 grams of aluminium cans. This is more meaningful, more valuable data
even though it is still a pile of 1kg of garbage in total. But with such preliminary analysis (sort, sift
and measure), that should prompt you what to do next recycle! And not simply buried them at a
landfill.
Consider
missing
values
Preliminary
Normality data Consider
test analysis outliers
(PDA)
Quality of
measurement model
And remember, the statistical software wont tell you which analyses to run for your data, so
it is important that you have a good understanding of general statistical principles, including which
analyses are appropriate for different types of data.
Consider case 74 in the following example. The respondent answered 5 for all questions
having 5-point Likert scale. This case was removed from the actual analysis.
When to use:
Copy the formula to the rest of the cases > find cases with 0 (zero) variance > delete the
case(s).
Example:
=VAR.S(G2:AI2)
Both missing values and outliers can happen by chance in any data distribution. While the
missing values are common in a large data set, the outliers may indicate measurement errors or that
the sample has a heavy-tailed distribution. In either case, the outliers can be removed or maintained
depending on context of the study and of course, with a sound theoretical backing.
In SPSS, removing the data is pretty straight forward press delete button or exclude the
case from any subsequent analysis. The earlier is simple but not recommended because the
case/data is completely removed. My preference is the latter option because that allows future
analyses or reviewers to work with the incomplete data. In the latter, SPSS maintains the data but
does not account for it in the statistical analysis.
When to use:
To run recode:
Select required variables > move to Variables field > click Old and New Values
Select System- or user-missing > New Value > Value > type 99 > Add > Continue
Variable view.
As for the outliers, you can delete the case containing the outliers as in step 1 above or
constraint the outliers to a maximum (or minimum) value. Put is simply, you replace the outlier data
with a value that is more likely, more meaningful based on a logical context of the study, and of
course, a sound theoretical backing.
1
var = the variable of your choice that contains missing values.
2
99 = my personal coding structure to instruct SPSS how to mark missing values in the data set. You can
specify any number on the Variable view > Missing value column > type the number of your choice.
9 Preliminary data analysis: An analysis before the analysis | SOAR Centre
Transfrom > Recode into different variables > select
variable(s) > > Old and New values
Select Range > key in the range (e.g. 6 through 98) >
New Values (e.g. 5) > Add > Continue > OK.
Although method 1 and 2 are the simplest, many studies do not recommend such
treatments. This leaves us with option 3 (MI) and 4 (EM) as viable alternatives.
Before running MVA, it is advisable to recode your missing values. In this example, the
missing values are coded as 99. Run Recode into Same Variables process.
Select all variables > move to Numeric Variables field > Old and New Values
To run MI:
Selects required variables > move to Variables in Model field > Set imputations values
(default is 5) > Location of Imputed Data > Create a new dataset > Dataset name: > Type MI
> OK
Once completed, you should have a new dataset with N x i + N number of cases, where N is
the number of cases in your original dataset, i is the number of imputations you specify in the MI
procedure. In our example, N = 131 (original dataset), N = 786 (131 x 5 + 131) (see Figure 1.9). By
default, SPSS adds a new variable called Imputation_ to identify several datasets in the file as follow:
Using the values after the 5th imputation, you can split this dataset as a new file.
In Select cases: If dialog > Click variable Imputation Number on the left panel > click move
icon to move the variable to right panel > click = sign > type 5 > Continue > OK. (see Figure
1.8)
Figure 1.7: Parameters settings to select 5th imputation dataset as a new file.
Another preferred method dealing with missing data is expected maximisation (EM)
technique. EM is an iterative procedure producing variances, covariances and means in an
initial step, and repeating the whole process until changes in the parameters are so small,
that the final solution is said to have converged (Graham, 2012). EM is argued to be more
suitable when the pattern of missing values is at least missing at random (MAR). However,
EM performs best when the missing data is missing completely at random (MCAR). For
discussions on MAR and MCAR, see Graham (2012), Karanja, Zaveri, and Ahmed (2013) and
Bennett (2001).
Figure 1.10) > Select required variables > move to Quantitative Variables field
Click button EM > Select Save completed data (see Figure 1.11) > Create a new
dataset > Dataset name: > Type PostEM > Continue > OK.
Look at Univariate Statistics table in the result. You should check for percentage of missing
data for each variable. Some scholars (e.g. Hair et al., 2006; Scheffer, 2002) set the missing value
upper limit to 20%. Anything more than this ceiling can potentially bias the final result.
The following table is the most important. This is a result of Littles MCAR test (Little, 1988).
In this case, the p-value is not significant3 (i.e. p > .05) indicating that the pattern of missing values in
the data set is MCAR. Thus, the imputed values derived from EM procedure can be used for
subsequent analyses.
3
Littles MCAR tests the pattern of missing values to be missing completely at random (MCAR). If the p-value is
not significant, that means the pattern of missing values in the dataset is similar to an expected MCAR pattern.
17 Preliminary data analysis: An analysis before the analysis | SOAR Centre
Table 1.2: Little's MCAR test.
You can also check for normality using z-score of kurtosis and skewness of each variable. Use this
formula to calculate for zk and zs.
Zk and Zs values should be within
2.58 for p = .01 or 1.96 for p =
.05 (Hair, Black, Babin, Anderson,
& Tatham, 2006) for the data to
be considered as normally
distributed.
4
A simple discussion is also available at following webpages
http://www.stattutorials.com/SPSS/TUTORIAL-SPSS-Assess-Normality.htm and
http://www.isixsigma.com/tools-templates/normality/tips-recognizing-and-transforming-non-normal-data/
18 Preliminary data analysis: An analysis before the analysis | SOAR Centre
Select variable comp skill from
left panel > move to Dependent
List
Box plot:
Potential outlier in the dataset
There are two ways to deal with non-normal data consider the non-parametric statistics
OR transform the data.
If you opt to proceed with stats that require normality, a data transformation is certainly
your next move. There are 5 different methods to transform non-normal data into an approximate
normal distribution (Osborne, 2010). These are:
5
A practical explanation is available at
http://www.isixsigma.com/tools-templates/normality/making-data-normal-using-box-cox-power-
transformation/
Should you wish to study the theoretical makeup of the Box-Cox transformation, the original article of Box and
Cox (1964) is available from the reference list of this document.
24 Preliminary data analysis: An analysis before the analysis | SOAR Centre
The first 4 in the list are all power transformation which means that mathematical square
and/or square roots are applied. Although the log transformation method is common in social
science (Osborne, 2010), Box-Cox transformation is a popular non-normal data transformation in
many studies. However, there is no straight-forward method for Box-Cox transformation in SPSS.
6
Figure 1.15: Reliability test result .
6
This reliability test was run on an actual research data, conducted in 2011, which was based on Technology
Acceptance Model (TAM). It comes to no surprise that the Cronbachs alpha is high because the items were
adapted from a pool of highly cited studies, not to mention that the TAM itself is a saturated researched
theory.
27 Preliminary data analysis: An analysis before the analysis | SOAR Centre
1.5.2 Content validity
When you have a pool of items designed to measure a construct then you should be able to
demonstrate that a great care has been made during the construction of the survey instrument. The
content validity is the degree to which the items represent the dimensions of the construct being
measured (Hu, Dinev, Hart, & Cooke, 2012). The validity of the items is normally assessed through
literature and also preferably reviewed by domain experts (Straub, Boudreau, & Gefen, 2004). The
instrument has to be carefully constructed and pilot tested.
Now it becomes obvious why the literature review and the proposal stage is pretty much
7
critical prior to the data collection. It is unconventional to have the items reviewed after the data
has been collected. In short, content validity is a subjective evaluation of the items which can be
accomplished through:
i. Literature
ii. Domain expert review
Weight Temperature
(Construct) (Construct)
Figure 1.16: The construct (latent factor or variable) is measured by items (observed variables).
7
Make a note of this statement. Well visit this later down the road.
28 Preliminary data analysis: An analysis before the analysis | SOAR Centre
The construct validity can be assessed by:
i. Convergent validity: AVE > .5, square root of AVE > variance between constructs
ii. Discriminant validity: MSV < AVE and ASV < AVE
The convergent validity is the extent to which the items correlate with each other within
their parent construct (factor). While the discriminant validity is the extent to which the items do not
correlate with other items of a different construct. In short, the convergent validity means that the
theoretically predicted items measuring a factor only load well into that factor; not outside its
parent factor. Conversely, the items which are predicted by theories to be not measuring a construct
must have low loading outside their parent factor.
In the bathroom scale and thermometer case kilograms and pounds should correlate well
with each other, measuring the weight, while Celsius and Fahrenheit correlate with each other
measuring the temperature. If kilogram correlates with Celsius, that suggests we have a large cross
loading value indicating kilogram is also measuring temperature (which doesnt make sense!). This
violates the theory and shows that we have a serious discriminant validity issue. If on the other
hand, Celsius and Fahrenheit do not correlate well, this indicates we have convergent validity
problems.
Figure 1.18: Discriminant validity analysis (an excerpt from Dinev, Goo, Hu, & Nam, 2009, pp. 403-404).
In many occasions, the non-response has nothing to do with human psychological reluctance
to reply, its just a bad sample. Its not a rocket science to understand that if you mail a
questionnaire to a company which has moved to a different address, then surely that contributes to
non-response bias. Therefore, in a case of mail survey, it is important to use the most current
database of addresses to reduce the non-response bias. This step also applies in the email survey.
After all the precautions you make, you must be wondering how many is considered
sufficient. The answer is theres no hard rule for that. REF Shannon (1948) put 70% response rate is
acceptable while REF Kerlinger (1968) agreed at 80%.
However, a little consolation was offered by Paxson (1995) who posited that low response
rate is inevitable and tolerable. This is because the studies in social science discipline mostly centre
on homogenous populations with a strong commonality in group identity that affect decision making
people in this homogenous group make a decision on a pretext of being a member of the group
rather than an isolated individual (Leslie, 1972). Leslies analysis of 29 studies showed that the non-
response bias only affects the demographic variables, rather than the independent variables, on the
dependent variables.
As a precautionary however, its a good and wise move to run statistical analyses to test for
the non-response bias. Scholars have been using proxies to test for the non-response bias because
technically its very hard, if not impossible, to get a response from those who choose not to respond.
Testing the presence of the non-response bias can be made by:
i. Small scale post-hoc analysis effectively, you run a small-scale data collection after
the actual data collection stage ends, and later compare if the results of the two to
see if theres any significant difference.
ii. Split the data between two response waves (early and late reply)8.
8
Use t-test procedure (Section 4.1.1) for each variable in the study to find if theres any difference. If your
dataset is not normally distributed, use Mann-Whitney U (section 6.1) to test differences among early and late
respondents.
31 Preliminary data analysis: An analysis before the analysis | SOAR Centre
1.6.2 Common method bias
Common method bias (CMB) is a measurement error (Podsakoff, MacKenzie, Lee, &
Podsakoff, 2003; Podsakoff, MacKenzie, & Podsakoff, 2012) that threatens the validity of a
conclusion drawn upon statistical results. This bias is observed via a presence of a systematic
variance (Bagozzi & Yi, 1990) that can inflate or deflate a given relationship among variables (Doty &
Glick, 1998) leading to unsound conclusions.
The common method bias can be attributed by raters (e.g. consistency motif and social
desirability), item charateristics (e.g. complex and ambiguous items) and context (e.g. context-
induced mood), and measurement context (e.g. time and location of measurement, common
medium to obtain measurement) (Podsakoff et al., 2003)9. Identifying the sources of the common
method bias allows us to better control for their influence on the data. Both procedural and
statistical measures are normally used to control for the biasing effect. The procedural measures
concern over the approach the data is collected and the instrument design. The measures include:
i. Obtain measures of the predictor and criterion variables from different sources.
ii. Use temporal, proximal, psychological or methodological separation of
measurement.
iii. Protect respondent anonymity and reduce evaluation apprehension.
iv. Counterbalance question order.
v. Improve scale items.
As much as the procedural methods are important, the statistical approaches also play
another objective role to control for the common method bias influence. Statistical controls against
the common method bias are:
Among all 4 main methods above, Harmans single factor test is the simplest measure. The
test is the most widely used in the literature (Podsakoff et al., 2003). What is required is a little
tweak in the factor analysis. Simply load all variables into the factor analysis, but constraint the
number of factor to 1. From the result, search for Total Variance Explained table and look at the
first component. If the first component accounts less than 50% of the all variables in the model, take
a deep breath and smile! Your instrument is free from significant common method bias effects.
The justification for the conclusion is that, if CMB is presence, than the unrotated factor
analysis will show that one item accounts for the majority of the variances in the model. Podsakoff
et al. (2003) disputed this approach claiming that the technique does nothing to control for the CMB
9
See Podasakoff (2003) and Bagozzi (1998) for a good discussion on the common method bias.
32 Preliminary data analysis: An analysis before the analysis | SOAR Centre
effect. Of course, it doesnt! Its a test to see if theres any substantial CMB presence in the data. The
only challenge you face using this method is to scratch your head looking for justifications when the
result shows you have more than 50% of the total variance explained by a single factor (see Figure
1.19).
Figure 1.21: Select all independent variables to test for the common method bias.
However, not all studies can be made in such contextual manner. Thus, concern over CMB
persists. Podsakoff et al. (2003) recommended a very assessment as to which method is suitable for
which study context. Figure 1.24 shows the summary of 7 study contexts in which different controls
can be made against CMB. It should be noted that the figure should be read together with Appendix
4.
10
Do you still remember footnote 7? Again, this is another good reason why the research proposal stage is
critical. Not only it defines your study context, but also the very mechanisms you are going to use for the data
collection that will substantially impact downstream stages data analysis, discussion and conclusion.
37 Preliminary data analysis: An analysis before the analysis | SOAR Centre
Figure 1.24: Controlling for common method bias (Podsakoff et al., 2003, p. 898).
1. Extraction methods:
a. Principal component analysis (PCA)
b. Principal axis factoring (PAF)
c. Maximum likelihood (ML)
When to use:
Descriptives > Select Initial solution > Select KMO and Barletts test of sphericity.
Example:
1. File: FactorAnalysis.sav
2. Find which items cluster into which factor or component.
11
Manifest variables, observed variables and indicators refer to the same type of variables, i.e. variables which
can be directly measured. These terminologies are used interchangeably throughout this text.
39 Factor analysis | SOAR Centre
Figure 2.1: Factor analysis > Descriptives settings.
Rotation > select Varimax > Display > Select Rotated solution.
Options > Exclude cases listwise > Coefficien Display Format > Select Supress small
coefficients > Absolute value below > Type .3.
Sig. .000
Depending on the extraction method, rotated solution is normally used to identify which
item belongs to which component. Some scholars use minimum loading of .5 with low crossloading
as a basis to evaluate items to be retained for composite scores. Note that in the following rotated
solution, loading of less than .3 is not displayed. This was supressed in the steps above.
Component
1 2 3 4 5 6
ease1 .796
ease2 .336 .704
ease3 .411 .731
ease4 .790 Internalisation and Identification items
ease5 .740 load into a single factor. Depending on
ease6 .711
the questions and theories, you may
want to re-specify these as a single factor
usefulness1 .648 .311 .394
rather than 2 distinct factors.
usefulness2 .655 .355
usefulness3 .799
usefulness4 .805
usefulness5 .824
usefulness6 .596 .362
Intention1 .779
Intention2 .322 .309 .755
Intention3 .786
Intention4 .726
Attitude1 .780
Attitude2 .837
Attitude3 .319 .806
Internalisation1 .787
Internalisation2 .797
Internalisation3 .766
Identification1 .725
Identification2 .394 .315 .609
Identification3 .403 .697
Compliance1 .708
Compliance2 .796
Compliance3 .849
Compliance4 .855
1. Summated scale. Add up raw scores from all indicators which load into a given factor.
2. Means. Calculate mean scores for all indicators for a given factor and sum them.
3. Factor scores. Add factor scores for all indicators for a given factor.
Of all three methods above, mean gives researchers more control over the calculations (Hair
et al., 2006) and facilitates interpretation of descriptive analysis results.
Example:
1. File: CompositeScore.sav
2. Find which items cluster into which factor or component.
4. A new variable
called PEOU is
added in the last
column.
4.1 t-test
This type of test provides a basis to analyse if there is any difference between two sample
groups, e.g. between male and female, before and after treatment (in an experiment), and
early and late respondents. If the test is significant (i.e. p < .05) then we can conclude that the
two samples means are statistically different.
In order to run independent samples t-test, the following 4 assumptions have to be met:
Define Groups > Group 1 > Type 1 > Group 2 > Type 2 > OK.
Example:
1. Dataset: IndependentTTest.sav
2. This exercise is to test any significant difference in attitude between male and female
participants toward e-learning system.
Continue > OK
In order to run related (paired)-samples t-test, the following 3 assumptions have to be met:
Select the variables to be tested > move to Paired Variables field > OK.
Example:
Shows whether
correlation
between before
and after course
groups is
significant.
When to use:
In order to run independent samples ANOVA, the following 4 assumptions have to be met:
Click Post Hoc > Equal Variances Assumed12 > Select Tukey > Continue
Example:
12
Tukey test is appropriate for equal sample size.
Gabriel procedure is more appropriate for an unequal sample size.
Hochberg GT2 is appropriate when the sample size is extremely unequal.
Games-Howell is normally when equal variance cannot be assumed (i.e. Levenes test is significant).
51 Test of differences among groups | SOAR Centre
52 Test of differences among groups | SOAR Centre
Extremely unequal sample
size. Read result from
Hochbers GT2.
Homogeneity of
variance
assumption holds.
When to use:
1. Linearity. The relationship of covariate and DV is linear. Scatterplot can be used to assess
this relationship.
2. Normality. Each latent variable is normally distributed.
3. Independent. Each participant is measured once and has no influence on others in the
survey or experiment.
4. Homogeneity of variance. Each set of scores is approximately equal. This is assessed by
Levenes test which is part of the ANOVA/ANCOVA result.
5. Homogeneity of regression slope. The regression slope of IV measured on DV has to be the
same for all groups.
To run ANCOVA:
In Univariate dialog:
> Move DV to Dependent Variable field
> Move IV to Fixed Factor(s) field
> Move covariate (e.g. gender) to Covariate(s) field
In order to run independent samples Pearsons r, the following 4 assumptions have to be met:
To run Pearsons r:
Example:
1. Dataset: Pearson.sav
2. This exercise is to test if attitude correlates with intention to use e-learning.
Statistics > select Estimates, Confidence intervals level (%): 95, Model fit, Part and partial
correlations, Collinearity diagnostics
Plots > *ZPRED > move to X (axis), *ZRESID > move to Y (axis) > select Normal probability
plot
Save > select Distances: Mahalanobis, and Cooks (this is to check for multivariate outliers)
Example:
INTENT
ATT
USEFUL
Changes in F value is
significant from zero.
PEOU
INTENT
ATT
USEFUL
Non-normal if points
substantially deviate from
the diagonal line.
Assessment of normality,
linearity and
homoscedasticity of
residuals.
An absence of clear
pattern of data spread
indicates all three
assumptions above are
met.
Statistics > select Estimates, Confidence intervals level (%): 95, Model fit, Part and partial
correlations, Collinearity diagnostics
Plots > *ZPRED > move to X (axis), *ZRESID > move to Y (axis) > select Normal probability
plot
Save > select Distances: Mahalanobis, and Cooks (this is to check for multivariate outliers)
Example:
Parameters for Statistics, Plots, Save and Options remain the same as in Multipl regression section.
PEOU
INTENT
ATT
USEFUL
COMPLY
6.1 Mann-Whitney U
When to use:
To run Mann-Whitney U:
Analyze > Nonparametric Tests > Legacy Dialogs > 2 Independent Samples
Select the variables to be tested > move to Test Variable List field
Select grouping variable (e.g. gender, early and late reply) > move to Grouping Variable
field
Define Groups > Group 1 > Type 1 (e.g. 1 is male group) > Group 2 > Type 2 (e.g. 2 is
female group) > Continue
Example:
1. Dataset: MannWhitney.sav
2. This exercise is to test any significant difference in attitude toward e-learning system among
male and female participants of an e-learning course.
To run Kruskal-Wallis:
Analyze > Nonparametric Tests > Legacy Dialogs > k Independent Samples
Select the variables to be tested > move to Test Variable List field
Select grouping variable (e.g. expertise level, age group) > move to Grouping Variable field
Define Range > Minimum 1 > Type 1 (e.g. 1 is a group with the lowest skill) > Maximum >
Type 4 (e.g. 4 is the maximum group level - professional) > Continue
Example:
1. Dataset: KruskalWallis.sav
2. This exercise is to test any significant difference in intention to use e-learning system among
instructors with different levels of computer skill.
Transform data
Cronbachs alpha1.1.1
Reliability1.1
Quality of
Validity &
measurement
reliability tests1 Composite reliability1.1.2
model
Content validity1.2
Address biases2
1.3 Domain expert1.2.1
Construct validity
2.2. Common (correlated) method bias (see Bagozzi & Yi, 1990; Podsakoff et al., 2003;
Podsakoff et al., 2012)
13
CR = Composite Reliability
AVE = Average Variance Extracted
MSV = Maximum Shared Squared Variance
ASV = Average Shared Squared Variance
78 Appendices | SOAR Centre
7.4 Appendix 4: Statistical remedies for CMB
Allen, P., & Bennett, K. (2010). PASW statistics by SPSS: A practical guide, version 18.0 (1 ed.).
Sydney: Cengage Learning Australia Pty Limited.
Bagozzi, R. P., & Yi, Y. (1990). Assessing method variance in multitrait-multimethod matrices: The
case of self-reported affect and perceptions at work. Journal of Applied Psychology, 75(5),
547-560. doi: 10.1037/0021-9010.75.5.547
Bennett, D. A. (2001). How can I deal with missing data in my study? Australian and New Zealand
Journal of Public Health, 25(5), 464-469. doi: 10.1111/j.1467-842X.2001.tb00294.x
Box, G. E. P., & Cox, D. R. (1964). An Analysis of Transformations. Journal of the Royal Statistical
Society. Series B (Methodological), 26(2), 211-252. doi: 10.2307/2984418
Dinev, T., Goo, J., Hu, Q., & Nam, K. (2009). User behaviour towards protective information
technologies: The role of national cultural differences. Information Systems Journal, 19(4),
391-412. doi: 10.1111/j.1365-2575.2007.00289.x
Doty, D. H., & Glick, W. H. (1998). Common methods bias: Does common methods variance really
bias results? Organizational Research Methods, 1(4), 374-406. doi:
10.1177/109442819814002
Hair, J. F., Black, W. C., Babin, B. J., & Anderson, R. E. (2010). Multivariate data analysis (7 ed.).
Upper Saddle River, NJ, USA: Prentice-Hall, Inc.
Hair, J. F., Black, W. C., Babin, B. J., Anderson, R. E., & Tatham, R. L. (2006). Multivariate data analysis
(6 ed.). New Jersey: Pearson Prentice Hall.
Hu, Q., Dinev, T., Hart, P., & Cooke, D. (2012). Managing Employee Compliance with Information
Security Policies: The Critical Role of Top Management and Organizational Culture*. Decision
Sciences, 43(4), 615-660. doi: 10.1111/j.1540-5915.2012.00361.x
Karanja, E., Zaveri, J., & Ahmed, A. (2013). How do MIS researchers handle missing data in survey-
based research: A content analysis approach. International Journal of Information
Management, 33(5), 734-751. doi: http://dx.doi.org/10.1016/j.ijinfomgt.2013.05.002
Leslie, L. L. (1972). Are high response rates essential to valid surveys? Social Science Research, 1(3),
323-334. doi: http://dx.doi.org/10.1016/0049-089X(72)90080-4
Little, R. J. A. (1988). A test of missing completely at random for multivariate data with missing
values. Journal of the American Statistical Association, 83(404), 1198-1202. doi:
10.2307/2290157
Paxson, M. C. (1995). Increasing survey response rates: Practical instructions from the total-design
method. Cornell Hotel and Restaurant Administration Quarterly, 36(4), 66-66.
Podsakoff, P. M., MacKenzie, S. B., Lee, J.-Y., & Podsakoff, N. P. (2003). Common method biases in
behavioral research: A critical review of the literature and recommended remedies. Journal
of Applied Psychology, 88(5), 879-903. doi: 10.1037/0021-9010.88.5.879
Podsakoff, P. M., MacKenzie, S. B., & Podsakoff, N. P. (2012). Sources of Method Bias in Social
Science Research and Recommendations on How to Control It. Annual Review of Psychology,
63(1), 539-569. doi: doi:10.1146/annurev-psych-120710-100452
Scheffer, J. (2002). Dealing with missing data. Research letters in the information and mathematical
sciences, 3(1), 153-160.
Straub, D., Boudreau, M.-C., & Gefen, D. (2004). Validation guidelines for IS positivist research.
Communications of the Association for Information Systems, 13, 380-427.
Yeh, K.-J. (2009). Reconceptualizing technology use and information system success: Developing and
testing a theoretically integrated model. (Doctor of Philosophy Doctorate thesis), University
of Texas, Arlington, Texas.