Abstract
This article addresses Rönkkö and Evermann’s criticisms of the partial least squares (PLS) approach to structural equation modeling. We contend that the alleged shortcomings of PLS are not due to problems with the technique, but instead to three problems with Rönkkö and Evermann’s study: (a) the adherence to the common factor model, (b) a very limited simulation designs, and (c) overstretched generalizations of their findings. Whereas Rönkkö and Evermann claim to be dispelling myths about PLS, they have in reality created new myths that we, in turn, debunk. By examining their claims, our article contributes to reestablishing a constructive discussion of the PLS method and its properties. We show that PLS does offer advantages for exploratory research and that it is a viable estimator for composite factor models. This can pose an interesting alternative if the common factor model does not hold. Therefore, we can conclude that PLS should continue to be used as an important statistical tool for management and organizational research, as well as other social science disciplines.
Partial least squares (PLS) path modeling is widely used not only in management research but also in virtually all social sciences disciplines (Hair, Sarstedt, Pieper, & Ringle, 2012; Hair, Sarstedt, Ringle, & Mena, 2012; Lee, Petter, Fayard, & Robinson, 2011; Peng & Lai, 2012; Ringle, Sarstedt, & Straub, 2012; Sosik, Kahai, & Piovoso, 2009). Bemoaning the low coverage of PLS in selected research methods journals, Rönkkö and Evermann (2013, hereafter R&E) sought to examine “statistical myths and urban legends surrounding the often-stated capabilities of the PLS method and its current use in management and organizational research” (R&E, p. 443).
Unfortunately, what could have been a useful, critical, and objective study on the characteristics of PLS turns out to be a polemic. PLS users are told by R&E that they “use PLS for purposes that it is not suitable for” (p. 443); that they misunderstand “the relative capabilities of PLS and the commonly used SEM estimators” (p. 443); and that they even “prefer PLS over SEM” because of “the publication bias in many fields for ‘positive’ results” (p. 443). According to R&E, PLS does not even deserve the (quality) designation of “structural equation modeling” and its use is “very difficult to justify” (p. 443). What follows is a sampling of their inflammatory statements about PLS: “the idea that PLS results can be used to validate a measurement model is a myth” (p. 438); “the PLS path estimates cannot be used in NHST [null hypothesis significance testing]” (p. 439); “the small-sample-size capabilities of PLS are a myth” (p. 442); “PLS does not have [the capability to] reveal patterns in the data” (p. 442); “PLS lacks diagnostic tools” (p. 442); “PLS cannot be used to test models” (p. 442); and “PLS is not an appropriate choice for early-stage theory development and testing” (p. 442).
But how do R&E actually arrive at these damning conclusions, which run contrary to many review studies across disciplines (e.g., Gefen, Rigdon, & Straub, 2011; Hair, Ringle, & Sarstedt, 2011; Hair, Sarstedt, Ringle, et al., 2012; Hulland, 1999; Peng & Lai, 2012; Sosik et al., 2009)? And how is it possible that R&E cannot find even a single positive attribute of PLS?
The main reason for the majority of R&E’s findings and conclusions lies in a central but undisclosed assumption they make: that the common factor model is indeed correct. The common factor model hypothesizes that the variance of a set of indicators can be perfectly explained by the existence of one unobserved variable (the common factor) and random error. R&E seem to hold such a deep belief in the undisputed tenability of the common factor model that they fail to grasp that virtually all of their findings only hold conditionally on the assumption that the common factor model is true. At the same time, they do not accept alternative approaches to structural equation modeling (SEM) that do not assume a common factor model, such as PLS.
In their urge to discredit PLS, R&E end up making up new myths, such as the statement that Dijkstra’s (1983) findings led “to abandonment of further development” of PLS (p. 425) or that “PLS has largely been ignored in research methods journals.” Both statements are patently untrue. Indeed, there has been substantial further development of PLS—including by Dijkstra himself1—and there is a large body of literature on PLS published in other methods and statistics journals than the ones considered by R&E, including Psychometrika (A. Tenenhaus & Tenenhaus, 2011; Dijkstra & Schermelleh-Engel, in press), Journal of Marketing Research (Fornell & Bookstein, 1982; Hwang, Malhotra, Kim, Tomiuk, & Hong, 2010; Jagpal, 1982), Multivariate Behavioral Research (Lohmöller, 1988; McDonald, 1996), and Computational Statistics & Data Analysis (Hanafi & Qannari, 2005; Jakobowicz & Derquenne, 2007; M. Tenenhaus, Esposito Vinzi, Chatelin, & Lauro, 2005).
In the remainder of this article, we present PLS as an SEM technique designed for estimating parameters of composite factor models and point out that the common factor model is nested in the composite factor model. Furthermore, we demonstrate that R&E’s conclusions about the characteristics of PLS are not justified and are partly an artifact of the setup of their simulation study. In fact, R&E inadvertently created six new myths, which we dispel. By reviewing and reconsidering R&E’s claims, our article contributes to reestablishing a constructive discussion of the PLS method and its properties.
Critique 1: Is PLS an SEM Method?
R&E argue that PLS is not truly an SEM method because it produces inconsistent and biased estimates and lacks an overidentification test. R&E’s claim that PLS is not SEM is novel and directly opposes the view of the original articles on the method. For example, Hui and Wold (1982) state that “during the recent interests among econometricians in the study of structural equation models and models with error variables, two lines of research have emerged: (a) Jöreskog’s maximum likelihood approach … and Wold’s distribution free Partial Least Squares (PLS) approach” (p. 119).2 Since the question of whether PLS is an SEM method goes beyond simple semantics, it requires careful examination. As R&E do not provide a definition of SEM,3 we examine their claim by analyzing what, according to extant literature, constitutes SEM.
To begin with, SEM is not a single technique, but “a collection of statistical techniques that allow a set of relations between one or more independent variables (IVs), either continuous or discrete, and one or more dependent variables (DVs), either continuous or discrete, to be examined” (Ullman, 2006, p. 35). It is “a synthesis of procedures developed in econometrics, and psychometrics” (Bollen & Long, 1993, p. 1). According to Byrne (1998),
The term structural equation modeling conveys two important aspects of the procedure: (a) that the causal processes under study are represented by a series of structural (i.e., regression) equations, and (b) that these structural relations can be modeled pictorially to enable a clearer conceptualization of the theory under study. (p. 3)
Importantly, any model represented as a structural equation model can also be expressed as a set of restrictions imposed on the covariance matrix of the observed variables. It is these restrictions (constraints) that make a structural equation model a model (Jöreskog & Sörbom, 1993). For instance, the common factor model restricts the covariance between two indicators of one construct to the product of their loadings and the construct’s variance (Bollen, 1989).
Many researchers (R&E apparently being no exception) mistakenly believe that constructs and common factors are one and the same thing (Rigdon, 2012). For them, SEM is a common factor-based technique, and constructs are identical to common factors. They thus follow a “fixed habit in applied research” (McDonald, 1996, p. 266) and fail to recognize that “structural equation models allow more general measurement models than traditional factor analytic structures” (Bollen & Long, 1993, p. 1). From a philosophical standpoint, there is no need for modeling constructs as common factors (e.g., Kaplan, 1946), and reducing SEM to common factor models is a very restrictive (unnecessarily restrictive, we would argue) view about SEM. Moreover, despite the parsimony, elegance, relatively easy interpretability, and widespread dissemination of the common factor model, scholars have started questioning the reflex-like application of common factor models (Rigdon, 2012, in press). A key reason for this skepticism is the overwhelming empirical evidence indicating that the common factor model rarely holds in applied research (as noted very early by Schönemann & Wang, 1972). For example, among 72 articles published during 2012 in what Atinc, Simmering, and Kroll (2012) consider the four leading management journals that tested one or more common factor model(s), fewer than 10% contained a common factor model that did not have to be rejected.4
A more general measurement model is the composite factor model. The composite factor model relaxes the strong assumption that all the covariation between a block of indicators is explained by a common factor. This means the composite factor model does not impose any restrictions on the covariances between indicators of the same construct. Instead, the name composite factor model is derived from the fact that composites are formed as linear combinations of their respective indicators. These composites serve as proxies for the scientific concept under investigation (Ketterlinus, Bookstein, Sampson, & Lamb, 1989; Maraun & Halpin, 2008; Rigdon, 2012; M. Tenenhaus, 2008).
Figure 1 provides a graphical representation of a composite factor model and the corresponding common factor model. Both models consist of two related constructs with three observed indicators each. The variances of the constructs are constrained to one. For the common factor model, the implied covariance matrix is then,
| 1 |
The illustrative common factor model has 13 parameters (six loadings λi, six residual (error) covariances θi, and one structural covariance φ), which is substantially fewer than the number of moments, which is 21. The example common factor model thus has 21 – 13 = 8 degrees of freedom.
The composite factor model leaves the covariation between indicators of the same block unexplained, which means that the implied covariances between these indicators equal the empirical covariances σij. This can also be expressed by additional parameters that capture the remaining indicator covariances that are not explained by their construct. Consequently, the composite factor model is less parsimonious than the common factor model. However, since the number of moments still exceeds the number of parameters (i.e., the number of degrees of freedom is greater than zero), we can refer to it as a model.5
A basic principle in soft modeling is that all information between the blocks is conveyed by the composite factors (Wold, 1982). Thus, although the residual covariance matrix is no longer diagonal, it is still block diagonal. The implied covariance matrix of our example composite factor model is then,
| 2 |
Comparing the two implied covariance matrices, it is evident that the common factor model is nested within the composite factor model because it has the same restrictions as the composite factor model plus some additional restrictions. The fact that the common factor model is nested within the composite factor model implies three possible combinations regarding model fit: (a) neither of the two models fits the data, (b) only the composite factor model fits the data, or (c) both models fit the data. This nestedness has important consequences for researchers’ choice of a model. In the first case, if neither of the two models fits the data, researchers should reject both models and refrain from interpreting estimates. In the second case, if only the composite factor model fits the data, researchers should use statistical techniques that are designed for composite factor models (e.g., PLS). In the third (less likely) case, if the common factor model fits the data, researchers should choose the common factor model and apply other appropriate estimation techniques (e.g., covariance-based SEM).
Composite factor models are likely far more prevalent than common factor models (Bentler & Huang, in press) and they are likely to have a higher overall model fit (Landis, Beal, & Tesluk, 2000). Moreover, the use of composite factor models in SEM occurs more often than one might expect. For example, the standard recipes to counter Heywood cases—fixing the error variance to zero or a small positive value—essentially abandon the common factor model for a mix between a common factor and a composite factor model (Bentler, 1976). Similarly, modifying an initial common factor model such that measurement errors are allowed to correlate coincides with a move toward a composite factor model.
Bearing the nature of the composite factor model in mind also helps better understand the nature of the alleged bias in PLS estimates. It is critical to first realize that the implied covariances between indicators of different constructs as determined by PLS are unbiased (Areskoug, 1982; Lohmöller, 1989). In other words, PLS estimates are not inherently biased, but only appear to be biased when interpreted as effects between common factors instead of effects between composite factors. Hence, bias cannot be a reason for PLS not to be seen as an SEM method. Moreover, while the lack of an overidentification test in PLS has repeatedly been criticized (e.g., Gefen et al., 2011; Hair et al., 2011), there is in fact no reason that prevents the testing of the discrepancy between the observed covariance matrix and PLS’s model-implied covariance matrix (and, indeed, we do this later when addressing Critique 3). We conclude, therefore, that PLS is clearly an SEM method that is designed to estimate composite factor models.
Critique 2: Are PLS Construct Scores More Reliable Than Sum Scores?
According to R&E, it is a myth that “PLS reduces the effect of measurement error” (p. 434). In fact, they claim that if anything, PLS provides “lower reliability composites than even simple summed scales” (p. 436). Neither statement is universally true and should therefore not be taken at face value.
While PLS does not completely eliminate the effects of measurement error, it does reduce it substantially.6 This is because component methods with multiple indicators achieve a degree of adjustment for unreliability, due to the very nature of weighted composites (e.g., Creager & Valentine, 1962; Ley, 1972; Mulaik, 2010; Rigdon, 2012). Let C be a composite of p weighted variables xi (i = 1, … p)
| 3 |
where the wi are the weights for multiplying each respective variable before adding it to the composite. Then, the variance of the composite C is given by
| 4 |
where is the variance of xi and (i = 1 to p, j = 1 to i, i ≠ j) the covariance between two different indicators xi and xj (e.g., Mulaik, 2010). Rigdon (2012) provides a perspective on this equation when he says,
On the surface, covariances count twice. However, the variances themselves can be divided into shared variance and unshared or unique variance. The unique variance of each indicator, then, is counted only once in the variance of the composite, while shared variance is counted three times—once within the variance term of the indicator and twice among the covariances. So unique variance, including random measurement error, is substantially underweighted in forming the composite. (p. 346)
Thus, the very act of creating a weighted sum helps account for measurement error. PLS further accentuates this effect in that it places more weight on the more reliable indicators. More precisely, PLS determines the indicator weights such that the corresponding construct yields a high predictive relevance (i.e., the construct fulfills its role in the nomological net as well as possible; see Rigdon, 2012; Sarstedt, Ringle, Henseler, & Hair, in press). But if this is true then, why did R&E’s analysis yield results that contradict this logic? The answer to this question lies in their extremely limited simulation study design.
A fundamental concern in any simulation study relates to the design of the research model as well as the choice of design factors and their levels. To ensure external validity of results, the design of any simulation study needs to closely resemble setups commonly encountered in applied research (Paxton, Curran, Bollen, Kirby, & Chen, 2001). This is particularly true for simulation studies comparing the performance of statistical methods aimed at providing researchers with recommendations regarding their use in common research settings (Mooney, 1997; Rubinstein & Kroese, 2008).
Research on PLS has generated a multitude of different simulation studies that compare the technique’s performance with that of other approaches to SEM (Areskoug, 1982; Goodhue, Lewis, & Thompson, 2006, 2012; Hulland, Ryan, & Rayner, 2010; Hwang et al., 2010; Lu, Kwan, Thomas, & Cedzynski, 2011; Reinartz, Haenlein, & Henseler, 2009; Vilares, Almeida, & Coelho, 2010; Vilares & Coelho, 2013). As can be seen from Table 1, these studies vary considerably in terms of their model setup. In this context and despite the fact that most recent simulation studies use quite complex models with a multitude of constructs and path relationships, R&E chose to use a two-construct model with a single path as their basis for their simulation.7 This, however, inevitably raises the question whether this model can indeed be considered representative of published research from an applied standpoint. As Paxton et al. (2001) note, when creating a model, “the Monte Carlo researcher should review structural equation model applications across a large number of journals in several areas of research to which they would like to generalize the subsequent results” (p. 291). Bearing this in mind, we revisited review studies on the use of PLS in marketing (Hair, Sarstedt, Pieper, et al., 2012), strategic management (Hair, Sarstedt, Ringle, et al., 2012), and information systems research (Ringle et al., 2012). Out of the 532 PLS models being estimated in 306 journal articles, there was exactly one model (0.2%) with two constructs by Rego (1998). More precisely, the average number of constructs was 7.94 in marketing, 7.5 in strategic management, and 8.12 in information systems, respectively. Therefore, R&E’s simulation model setup is not remotely representative of research studies using PLS.
|
Table 1. Prior Simulation Studies on PLS’ Performance.

In addition to the model setup, researchers need to carefully decide on the relevant factors and their levels. Comparing R&E’s simulation study with prior research in terms of design factors (see Table 1) further exposes the very limited scope of their study. Specifically, R&E vary neither the factor loadings nor the sample size, “one of the most important variables in a simulation” (Paxton et al., 2001, p. 294) and of particular relevance to PLS due to its well-known consistency at large property (e.g., Hui & Wold, 1982). The only factor that R&E vary is the standardized path coefficient between the two constructs across the two levels of 0 and 0.3, implying R2 values of 0 and 0.09. However, Paxton et al. (2001) recommend “that the R2s take values that are representative of the areas of research with which one is most concerned. For instance, R2 values with cross-sectional, individual-level data frequently range between 0.2 and 0.8” (p. 297). This is also true for PLS research in, for example, strategic management where the average R2 in the studies reviewed by Hair, Sarstedt, Pieper, et al. (2012) was 0.36.
Against this background, we first replicate R&E’s setup by estimating a two-construct model with 100 observations, indicator loadings of 0.6, 0.7, and 0.8, and two conditions for the structural model coefficient (Condition a: β = 0; Condition b: β = 0.3). Next, we add a condition in which the standardized path coefficient has a value of 0.5, implying an R2 value more commonly encountered in applied research of .25 (Condition c). In a third step, we examine how an increased sample size of 500 observations—as routinely assumed in simulation studies in SEM in general (Paxton et al., 2001) and PLS in particular (Hulland et al., 2010; Hwang et al., 2010; Reinartz et al., 2009; Vilares & Coelho, 2013)—would influence the findings (Condition d). In a fourth step, and in line with the results of prior PLS reviews (Hair, Sarstedt, Pieper, et al., 2012; Hair, Sarstedt, Ringle, et al., 2012; Ringle et al., 2012), we consider a more heterogeneous loading pattern (0.5/0.7/0.9; Condition e). Last, to capture the typically larger complexity of models, our simulation also includes a population model with four constructs (Condition f) as shown in Figure 2.8
Figure 3 provides a graphical representation of the reliabilities of construct A as obtained through our Monte Carlo simulation. The results of our replication in Panels a and b are consistent with R&E’s findings. For both PLS modes, sum scores provide better (more reliable) construct scores, and PLS Mode B performs substantially worse than Mode A. However, as can also be seen in Figure 3, a different picture emerges when we vary one or more factor levels. For example, when increasing the effect size between the two constructs to a moderate (β = 0.5) level, PLS Mode A most times yields more reliable construct scores than sum scores. Likewise, although still not fully matching up with PLS Mode A or sum scores, PLS Mode B performs considerably better than when assuming no effect (β = 0) or a small effect (β = 0.3). An increase of sample size also decreases differences between the three estimation modes, with PLS Mode A outperforming sum scores. The performance of PLS Mode A relative to sum scores improves even more when assuming a more heterogeneous set of indicators. Specifically, when assuming indicator loadings of 0.5, 0.7, and 0.9, both PLS Mode A and Mode B perform considerably better than in R&E’s initial setup. This result is not surprising since—as noted earlier—PLS prioritizes indicators according to their predictive validity; that is, those indicators with a smaller error variance contribute more strongly to the measurement of the latent variables (e.g., Henseler, Ringle, & Sinkovics, 2009). Last, embedding the latent variables in a broader nomological net also improves the performance of both PLS Modes, with Mode A again outperforming sum scores (Mode B is again dominated by the other estimation techniques).
Figure 3 also includes the reliability of the indicator with the highest loading. This ‘best’ indicator serves as a naïve benchmark for PLS’s capability of reducing measurement error. It is evident that construct scores obtained through PLS Mode A generally outperform the best item. Not surprisingly, the only exception is Condition e, where one indicator has a particularly high reliability (0.9).
Table 2 provides an overview of the average reliability scores of PLS Mode A, sum scores, and the best indicator estimation, and it indicates the percentage of simulation runs where PLS Mode A provided better (i.e., more reliable) construct scores.9 PLS Mode A outperforms the best indicator across all model constellations, providing support for the capability of PLS to reduce measurement error. Furthermore, only in the two model constellations examined by R&E does PLS Mode A lag behind sum scores in terms of reliability. In contrast, in all other, likely much more realistic settings (e.g., Hair, Sarstedt, Pieper, et al., 2012; Hair, Sarstedt, Ringle, et al., 2012; Ringle et al., 2012), PLS Mode A clearly outperforms sum scores.
|
Table 2. Simulation Results: Reliability of PLS Mode A, Sum Scales, and Best Indicator.

We conclude that PLS reduces measurement error but does not eliminate it. PLS should be preferred over regression on sum scores if there is sufficient information available to let PLS estimate construct weights under Mode A. Obviously, PLS construct scores can only be better than sum scores if the indicators vary in terms of the strength of relationship with their underlying construct. If they do not vary, any method that assumes equally weighted indicators will outperform PLS. We thus conclude that if the indicators differ in quality and PLS has sufficient information to estimate different weights under Mode A, the resulting PLS construct scores will be more reliable than sum scores.
Critique 3: Can PLS Be Used to Validate Measurement Models?
R&E examined to what extent selected model evaluation criteria help researchers detect measurement model misspecification. Based on a simulation study with four population models, they found that the selected evaluation criteria did not reliably distinguish between correctly specified and misspecified models. R&E thus concluded that “the idea that PLS results can be used to validate a measurement model is a myth” (p. 438).
Again, the conclusion reached by R&E must be viewed with great skepticism. Not only did they mistakenly assume that PLS estimates a common factor model rather than a composite factor model, they also committed several calculation and reporting errors and claimed—without any supporting empirical evidence—that covariance-based SEM would have a better performance in detecting model misspecification than PLS. More specifically, R&E mixed up two models in their reporting (the values for Model 2 were reported under Model 3 in their Table 4 and vice versa). They also calculated four out of five statistics incorrectly. First, they calculated both the composite reliability as well as the average variance extracted (AVE) assuming that the second population model (with only one construct) was estimated. Instead, R&E should have calculated values for each of the two constructs—further analysis typically ensuring that the minimum of the two AVE values is above 0.5. As a result, this Fornell–Larcker criterion is also miscalculated since it builds on the AVE statistic. Moreover, R&E did not compute the standardized root mean square residual (SRMR) as the root mean square discrepancy between the observed correlations and the model-implied correlations (Hu & Bentler, 1998, 1999), but the root mean square residual covariance (RMStheta) as suggested by Lohmöller (1989, Equation 2.118). Put simply, the vast majority of the reported values in R&E’s Table 4 are simply wrong. Finally, R&E did not assess the exact model fit of the PLS path model, so it remains unclear whether the PLS estimates can truly be used for model testing.
To verify and where necessary to correct R&E’s findings, we replicated their simulation study. Figure 4 shows the four population models used by R&E. For each population model, we generated 500 data sets with 100 observations each, and estimated the parameters using the first model, no matter by which population model the data were created.10 Analogous to R&E, we examine Jöreskog’s rho (as a measure of composite reliability), the AVE, and the Fornell–Larcker criterion (Fornell & Bookstein, 1982; Fornell & Larcker, 1981) as well as the relative GoF (GoFrel; Esposito Vinzi, Trinchera, & Amato, 2010) and the RMStheta (Lohmöller, 1989). Since R&E apparently intended to examine the SRMR, but actually did not, we also include this statistic in our analysis. We determine the SRMR for PLS in congruence with the model that is actually estimated, that is, an SRMR for a composite factor model.11 The model-implied covariance-matrix of the PLS path model is computed using Equation 2, and we determine the exact fit of the composite factor model by means of bootstrapping the conventional likelihood function. In essence, this constitutes a confirmatory composite analysis. Finally, to facilitate a comparison between PLS and covariance-based SEM, we also report the exact fit and the SRMR as obtained from covariance-based SEM. Table 3 reports the results of R&E and contrasts them with our own findings.
|
Table 3. Simulation Results: Performance of Different Indices of PLS and Covariance-Based SEM (CB-SEM) for Measurement Model Validation.

Model 1 represents a case in which the estimated model is isomorphic to the population model. Optimally, all assessment criteria should indicate that the model is acceptable. We find that this is the case for all assessment criteria other than the relative GoF and RMStheta. With an acceptance rate of 35.8%, the relative GoF is not really able to recognize that the estimated model is acceptable. This finding is consistent with Henseler and Sarstedt (2013), who likewise report that the family of GoF indices is not suitable for model comparisons and for detecting model misspecification. In fact, numerous articles clearly advise against the use of the GoF indices to validate PLS path models globally (Hair, Hult, Ringle, & Sarstedt, 2014; Hair, Ringle, & Sarstedt, 2013; Hair, Sarstedt, Pieper, et al., 2012; Hair, Sarstedt, Ringle, et al., 2012), which makes R&E’s inclusion of this statistic highly questionable.12 The only other criterion, which (erroneously) rejected Model 1 in R&E’s study, is RMStheta (which, as noted, R&E mistakenly labeled as SRMR). However, the threshold used for SRMR cannot be readily transferred to RMStheta. Lohmöller (1989) suggests only that the outer model is satisfactory “if the residual covariances between blocks are near zero” (p. 55).
Model 2 represents a situation in which an overparameterized model is estimated. Instead of assuming (and thus constraining) a correlation of one between the two constructs, the interconstruct correlation is freely estimated. In accordance with R&E, we find that none of the assessment criteria recommended for use in combination with PLS detect this lack of discriminant validity. This is particularly surprising with regard to the Fornell–Larcker criterion (Fornell & Larcker, 1981), which is intended exactly for conditions like the present one. However, we further note that the SRMRs of both PLS and covariance-based SEM as well as the relevant exact tests of fit likewise do not detect that the model is misspecified. While the acceptance rate of covariance-based SEM is somewhat lower than that of PLS (93.2% vs. 95.6%), it is obvious that neither PLS nor covariance-based SEM can reliably detect this form of misspecification.
Model 3 represents a situation in which two indicators are assigned to the wrong construct. While the AVE actually performs somewhat better than initially reported by R&E, the conventional assessment criteria of PLS do not reliably detect the misspecification. In contrast, the exact fit test and the SRMR of the composite factor model perform substantially better with acceptance rates of 1.4% and 3.0%, respectively. While, at first, the acceptance rate of PLS appears considerably worse than that of covariance-based SEM, the large number of Heywood cases—negative variances occur in more than 60% of the simulation runs—puts this into perspective. Ultimately, our overall evidence indicates that PLS provides a more reliable indication of model misspecification than covariance-based SEM.
Model 4 represents a situation in which the indicators are actually evoked by three constructs instead of two. Under this condition, the conventional PLS assessment criteria exhibit a better performance than previously. In particular, the AVE detects this form of misspecification reliably. Moreover, composite reliability performs relatively well. Again, the most reliable indicators of model misspecification are the exact fit and the SRMR. In particular, the test of exact fit rejects Model 4 in all cases. As in Model 3, the high number of Heywood cases (up to 100%, depending on the software implementation) renders the use of covariance-based SEM impracticable and puts its acceptance rate into perspective.
We conclude that while R&E’s critique on the efficacy of conventional measurement model assessment criteria (composite reliability, AVE, and the Fornell–Larcker criterion) is justified, their central conclusion is not. Researchers can rely on PLS-based assessment criteria such as the test of exact fit (i.e., statistical inference on the discrepancy between the empirical covariance matrix and the covariance matrix implied by the composite factor model) or the SRMR to determine to what extent the composite factor model fits the data. An additional criterion may be the RMStheta, which, in our study, was substantially higher for misspecified models than for correct models. Future research should try to identify an optimal threshold value for RMStheta to give researcher an additional instrument to recognize misspecified PLS path models.13 We conclude that PLS can help detect a wide spectrum of measurement model misspecifications as long as a composite factor model is assumed and the test of exact fit and/or the SRMR are used for model validation purposes.
Critique 4: Can PLS Be Used for Null Hypothesis Significance Testing (NHST) of Path Coefficients?
Based on a simulation study, R&E conclude that, because the distribution of parameter estimates is not normal but bimodal in shape when no effect between the two constructs under consideration is assumed (i.e., β = 0), PLS cannot be used in null hypothesis significance testing (NHST). Furthermore, they recommend that “bootstrapped confidence intervals of PLS estimates should not be used for making statistical inferences until further research is available to show under which conditions, if any, the bootstrapped distribution follows the sampling distribution of the PLS parameter estimate” (p. 440). Considering that PLS is routinely used for testing relationships derived from formal hypotheses (Rigdon, 2013) and that prior simulation studies have underlined PLS’s suitability for hypothesis testing across a wide range of model setups (e.g., Reinartz et al., 2009), the claim that PLS cannot be used in NHST is certainly disturbing. But is it true?
R&E’s claim is based on two arguments, namely that (a) the distribution of parameter estimates is not normal but bimodal in shape and (b) the difference between the distribution of parameter estimates and the empirical bootstrap distribution renders NHST impossible. Both arguments stand on quicksand. Regarding (a), R&E only considered one population model, but generalize to all models with path coefficients of zero. It remains thus to be seen whether the problem of bimodal distributions applies to all path coefficients of zero. Regarding (b), R&E’s argument is even shakier because they did not empirically investigate the robustness of NHST with respect to differences in distribution.
To shed further light on these issues, we replicate and extend R&E’s simulation design in a manner analogous to that for Critique 2 earlier. We first examine whether the bimodal shape of the parameter estimate distribution holds for other settings. What did we find? While our results shown in Figure 5 parallel those of R&E for a two-construct model with 100 observations, indicator loadings of 0.6, 0.7, and 0.8, and two conditions for the structural model coefficients (β = 0 and β = 0.3), this is not the case when the simulation design is varied slightly. More precisely, when increasing the standardized path coefficient to a moderate level (β = 0.5) or increasing the sample size (N = 500) or embedding the latent variables in a nomological net with constant effects (see Figure 2), parameter estimates clearly follow a unimodal distribution. Not surprisingly, assuming more heterogeneous loadings does not influence the shape of the distribution, which is in line with Reinartz et al.’s (2009) findings. In addition, the results show that the parameter bias is consistently smaller in PLS Mode A and Mode B estimation as compared to sum scores, making PLS more appealing from an accuracy perspective.
Next, since R&E question the suitability of PLS for NHST without actually examining its behavior, we analyze the Type I and Type II errors that PLS produces for R&E’s population model. We use the boot package (Canty & Ripley, 2013; Davison & Hinkley, 1997) to compare four ways of determining bootstrap confidence intervals (DiCiccio & Efron, 1996): (a) normal confidence intervals, which use bootstrap estimates of standard errors and quantiles of the normal distribution; (b) percentile confidence intervals, which rely on the percentiles of the empirical bootstrap distribution to determine confidence intervals around the parameters of interest; (c) bias-corrected and accelerated confidence intervals (BCa), which employ improved percentile intervals; and (d) basic bootstrap confidence intervals, as directly derived from the bootstrap percentiles. For a more detailed description, see Davison and Hinkley (1997).
In accordance with our earlier simulation design, we vary the path coefficient, sample size, and model complexity. Table 4 shows the results of our analysis, which draws on 500 replications. Three ways of determining bootstrap confidence intervals (normal, percentile, and BCa) largely maintain the prespecified alpha protection level of 5%, whereas the basic confidence interval produces an undue number of Type I errors. With regard to Type II errors, the BCa confidence intervals turn out to show the lowest power. However, BCa confidence intervals are the only ones that under all conditions maintained the alpha protection level of 5%. Overall, the normal, the percentile, and the BCa bootstrap confidence intervals appear suitable for NHST. Among them, the percentile bootstrap confidence interval is a good compromise (Hayes & Scharkow, 2013). These results are completely in line with those of Goodhue et al. (2012), who also did not find any problematic behavior of PLS in combination with bootstrapping.
|
Table 4. Simulation Results for Different Bootstrap Confidence Interval Types.

We conclude that PLS is generally suitable for NHST, even if the interrelated constructs are not embedded in a wider nomological net, if sample size is relatively low, and if expected effects are small. Researchers need to take care only that they use normal, percentile, or BCa bootstrapping, but not basic bootstrapping. Given that the dominant PLS software implementations of SmartPLS (Ringle, Wende, & Will, 2005) and PLS-Graph (Chin & Frye, 2003) already use normal bootstrapping (Temme, Kreis, & Hildebrandt, 2010), it is unlikely that researchers using this software have faced or will face problems with NHST.
Critique 5: Does PLS Have Minimal Requirements on Sample Size?
The most prominent argument for choosing PLS-SEM in many disciplines such as marketing (Hair, Sarstedt, Ringle, et al., 2012), management information systems (Ringle et al., 2012) and strategic management (Hair, Sarstedt, Pieper, et al., 2012) is the use of small sample sizes. This topic has been passionately debated in recent years (e.g., Marcoulides & Saunders, 2006) and has been empirically examined in various simulation studies (e.g., Areskoug, 1982; Goodhue et al., 2012; Hulland et al., 2010; Lu et al., 2011; Reinartz et al., 2009; Vilares & Coelho, 2013). As also emphasized in previous studies (e.g., Hair et al., 2013; Hair, Sarstedt, Pieper, et al., 2012; Hair, Sarstedt, Ringle, et al., 2012), we agree with R&E’s criticism that many authors seem to believe that sample size considerations do not play a role in the application of PLS by scholars.
Where does this idea come from then? We believe that this idea is fostered by the often cited “10 times” rule (Barclay, Higgins, & Thompson, 1995), according to which the sample size should be equal to the larger of (a) 10 times the index with the largest number of formative indicators or (b) 10 times the largest number of structural paths directed at a particular latent variable in the structural model. The argument becomes clearer if one notes that the role of sample size for SEM is twofold. On one hand, sample size is a major determinant of statistical power and therefore influences the quality of inference statistics obtained from a statistical technique (Paxton et al., 2001). On the other hand, each technique requires a certain sample size to be able to provide estimates (Hair, Anderson, Tatham, & Black, 1998). We examine both roles of sample size for SEM.
With regard to statistical power, it can be expected that covariance-based SEM, as a full-information estimator, will most of the time deliver smaller standard errors than limited-information estimators such as PLS. Empirical evidence based on simulation studies confirms this notion (Goodhue et al., 2012). Findings to the contrary—such as those by Lu et al. (2011) or Reinartz et al. (2009)—can be traced back to the differences we note above between the common factor model and the composite factor model. There are situations (population covariance matrices) in which there are effects between composite factors but not between common factors and vice versa. Depending on the population model, researchers can get the erroneous impression that one method has a higher statistical power to detect a certain effect than the other, whereas in fact the effects need to be interpreted differently (Rai, Goodhue, Henseler, & Thompson, 2013).
However, sample size not only plays a role with regard to statistical power. As Reinartz et al. (2009) show, PLS demonstrates better convergence behavior in the case of small sample sizes than covariance-based SEM. Our simulations confirm this finding. For instance, the sem package (Fox, Nie, & Byrnes, 2013) produced inadmissible solutions for all the 500 data sets drawn from the population 3 of Figure 4 (nonconvergence: 13.4%; Heywood cases: 86.4%). To verify that this poor behavior is attributable to the method itself and not a particular implementation of it, we triangulated this finding using the lavaan package version 0.5-14 (Rosseel, 2012), which produced inadmissible solutions in 63.4% of all cases (nonconvergence: 12.6%; Heywood cases: 50.8%) and Mplus version 4.21 (Muthén & Muthén, 2006), which resulted in inadmissible solutions in 70.0% of the cases (nonconvergence: 8.0%; Heywood cases: 62.0%). In contrast to covariance-based SEM, PLS always converged and always produced admissible solutions. These results are in accordance with extant research, which has found that nonconvergence and Heywood cases occur quite frequently with common factor models (Krijnen, Dijkstra, & Gill, 1998) and that PLS in general only rarely has convergence problems (Henseler, 2010).
Moreover, in contrast to covariance-based SEM, PLS can even be used if the number of observations is smaller than the number of variables (whether manifest or latent) or the number of parameters in the model. We therefore conclude that PLS can be applied in many instances of small samples when other methods fail.
Critique 6: Is PLS Appropriate for Exploratory or Early Stage Research?
R&E argue that “using PLS as an exploratory or early-stage theory testing tool does not feature strongly in the early PLS articles” (p. 442) and that only recent literature such as Hair et al. (2011) conveys this notion. This description openly misrepresents research reality. In fact, Herman Wold, the inventor of PLS, emphasized exploratory nature of PLS from the very beginning of its development:
PLS is primarily intended for research contexts that are simultaneously data-rich and theory-skeletal. The model building is then an evolutionary process, a dialog between the investigator and the computer. In the process, the model extracts fresh knowledge from the data, thereby putting flesh on the theoretical bones. At each step PLS rests content with consistency of the unknowns. As the model develops, it may be appropriate to try ML estimation with its higher aspiration of optimal accuracy. (Lohmöller & Wold, 1980, p. 1)
Moreover, as Table 5 clearly illustrates, Wold stressed this discovery-oriented process throughout the development phases of PLS (e.g., Wold, 1974, 1979, 1980, 1982, 1985). Rather than commit a priori to a specific model, he imagined a researcher estimating numerous models in the course of learning something about the data and about the phenomena underlying the data (Rigdon, 2013). Thus, PLS’s exploratory nature is certainly no “urban legend” but one of the driving forces of Wold’s PLS development.14
|
Table 5. Quotes by Hermann Wold Regarding the Intended Use of PLS.

While the originators of the PLS method stressed its exploratory character, this is usually not the case with applied researchers. Literature reviews show that only 43 of 306 articles (14.05%) published in marketing, strategic management, and information systems literatures mentioned “exploration” as a rationale for using PLS (Hair, Sarstedt, Pieper, et al., 2012; Hair, Sarstedt, Ringle, et al., 2012; Ringle et al., 2012). This is not all that surprising in light of the academic bias in favor of findings presented in confirmatory terms (Greenwald, Pratkanis, Leippe, & Baumgardner, 1986; Rigdon, 2013). In fact, this observation also holds for covariance-based SEM, which is traditionally viewed as a confirmatory tool. However, its actual use is far from it as researchers routinely engage in specification searches when the initially hypothesized model exhibits inadequate fit. Interestingly, this is more often the rule than the exception (e.g., Cliff, 1983; Sarstedt et al., in press). But does PLS have the necessary requisites for exploratory modeling? Contradicting the view of the originators of PLS, R&E argue that it does not, primarily because PLS cannot be used to validate models. R&E state that whereas covariance-based SEM can detect both overparameterization (by assessing the significance of path coefficients) and underparameterization (using modification indices), PLS can detect only overparameterization. Based on this observation, they conclude that PLS is not suitable for exploratory purposes. We do not agree with this conclusion for three reasons. First, R&E neglect an obvious modeling option: starting the exploratory analysis with a likely overparameterized or even saturated model and dropping nonsignificant paths. Second, as our answer to Critique 3 demonstrated, PLS can certainly be used to validate models, and at least two overall model criteria can detect underparameterization: (a) exact fit and (b) SRMR. Third, R&E fail to empirically demonstrate that covariance-based SEM performs better than PLS in finding the true model. Given the mediocre-to-poor performance of covariance-based SEM specification searches (e.g., Heene, Hilbert, Freudenthaler, & Bühner, 2012; Homburg & Dobratz, 1992), a blind faith in covariance-based SEM’s exploratory capabilities appears both unwise and unjustified.
R&E’s failure to clearly distinguish between the common factor model and the composite factor model also has consequences for their conclusions with respect to PLS’s adequacy for exploratory research. Owing to their blind adherence to the common factor model, R&E misinterpret Lohmöller: “The corollary assumed by Lohmöller, neither logically implied nor correct, is that PLS is appropriate when the researcher is not sure that the model is correct. Lohmöller [implies] that the term explorative refers to situations where the model may be incorrect” (p. 442). If things were really this way, R&E would be justified because model misspecification cannot be corrected by choosing one estimation technique over another (Henseler, 2012). Unfortunately, R&E ignore the fact that Lohmöller refers not to the model (as if there was only one model), but to the common factor model. We have demonstrated that the composite factor model is a less restrictive model than the common factor model. Therefore, Lohmöller’s notion that the composite factor model (as estimated by PLS) is appropriate when the researcher is unsure that the common factor model is correct has obvious merits. Similarly, one should understand Lohmöller’s use of the term explorative: Explorative latent variable path modeling refers to situations where the common factor model may be incorrect.
R&E also discuss a dichotomy of correct versus wrong models without acknowledging that models can be wrong in some subparts only. Particularly in early phases of research, it is difficult to assure that all subparts of a model are entirely correct. Full-information approaches often suffer in such instances because model misspecification in a subpart of a model can have detrimental effects on the rest of the model (Antonakis, Bendahan, Jacquart, & Lalive, 2010). In contrast, limited information methods are more robust to misspecification and are therefore useful for the analysis of initially formulated but misspecified models (Gerbing & Hamilton, 1994). Consequently, there are good reasons to prefer PLS as a limited-information approach over full-information approaches when the correctness of all parts of a model cannot be ensured.
Since PLS estimates a more general model than covariance-based SEM and is less affected by model misspecification in some subparts of the model, it is a suitable tool for exploratory research. We conclude that PLS can be a valuable tool for exploratory research.
Conclusion
R&E issued six critiques about PLS, which are covered by the following questions: (1) Is PLS an SEM method? (2) Are PLS construct scores more reliable than sum scores? (3) Can PLS be used to validate measurement models? (4) Can PLS be used for NHST of path coefficients? (5) Does PLS have minimal requirements on sample size? and (6) Is PLS appropriate for exploratory or early stage research? Their answer to all these questions is no.
Unlike R&E, we feel that the above questions require more nuanced answers than a simple yes or no. We provide such answers in Table 6 based on our arguments and simulation results.
|
Table 6. Summary of Conclusions.

The true value of R&E’s article lies not in the answers to the questions it poses but in pinpointing boundary conditions for the use of PLS and detecting weaknesses in some model assessment criteria. At the same time, the contribution of R&E to our understanding of PLS is counterproductive because of the strong assumption that the common factor model is (necessarily) correct, because extant literature was misread or misquoted, and because the results of specific population models were generalized too far. While trying to chase down myths, we argue that R&E have actually created new myths.
There is no such thing as an estimation method that is best for every model, every distribution, every set of parameter values, and every sample size. For all methods, no matter how impressive their pedigree (maximum likelihood being no exception), one can find situations where they do not work as advertised. One can always construct a setup where a given method, any method, “fails.” A (very) small sample or parameter values close to critical boundaries or distributions that are very skewed, thick tailed, and so on, or any combination thereof will do the trick. It is just a matter of perseverance to find something that it is universally “wrong.” A constructive attitude, one that aims to ascertain when methods work well, how they can be improved, and where they complement each other would seem to be more conducive to furthering the scientific enterprise.
An objective critique of any method should not only focus on its limitations but also highlight its advantages. R&E’s analysis falls short here in that it does not mention any of those characteristics that make PLS attractive for researchers. There are many more important questions about the behavior of PLS that deserve scholarly attention, such as these: “How well can PLS predict?” and “How well does PLS perform if the composite factor model is indeed correct?” Preliminary findings give rise to optimism for both these questions: PLS serves well for predictive purposes (Becker, Rai, & Rigdon, 2013), and when the composite factor model is true, PLS clearly outperforms covariance-based SEM and regression on sum scores (Rai et al., 2013). In conclusion, PLS is not a panacea, but certainly an important technique deserving a prominent place in any empirical researcher’s statistical toolbox.
Authors’ Note
The formal development of the composite factor model is attributable to the second author, and in collaboration with the first author developed into the concept of confirmatory composite analysis. The authors thank Dr. Rainer Schlittgen, Professor of Statistics and Econometrics, University of Hamburg, Germany, for helpful comments on earlier versions of the article.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Notes
1.
See Dijkstra (2010, in press), Dijkstra and Henseler (2012), or Dijkstra and Schermelleh-Engel (in press). Also ignoring Lohmöller’s (1989) seminal work does not add to the credibility of R&E’s arguments either.
2.
Karl Jöreskog and Herman Wold referred to “the ML [maximum likelihood] and PLS approaches to path models with latent variables” (Jöreskog & Wold, 1982, p. 270).
3.
Implicitly, the authors appear to equate SEM with a model for latent variables and with an estimation method that they require to be consistent and asymptotically normal (CAN). These joint demands are very unusual. Since PLS, as originally formulated and implemented, was not CAN, the accusation of “misleading” finds its justification only in their peculiar interpretation of SEM.
4.
The four leading management journals are Academy of Management Journal, Journal of Applied Psychology, Journal of Management, and Strategic Management Journal.
5.
By contrast, researchers thinking in terms of covariance-based SEM would typically not consider a linear regression among observed variables to be a model because it has zero degrees of freedom (e.g., Jöreskog & Sörbom, 1993).
6.
Note that a recent extension to PLS, termed “consistent PLS,” eliminates the effects of random measurement error (PLSc; see Dijkstra, in press; Dijkstra & Henseler, 2012; Dijkstra & Schermelleh-Engel, in press).
7.
As can be seen from Table 1, only the very first study in the field (Areskoug, 1982) uses as simple a simulation design as R&E do.
8.
Note that the aim of these extensions is not to provide a full-blown analysis of the effect of measurement error on PLS results but rather to show the influence of slight changes in the simulation design on the resultant findings.
9.
Because of its well recognized inferior overall performance, PLS Mode B is excluded from this overview.
10.
We used R 3.0.1 (R Core Team, 2013) to generate the data and aggregate the results, the semPLS package (Version 1.0-10; Monecke & Leisch, 2012) to estimate the composite factor model, and the sem package (Version 3.1-3; Fox, Nie, & Byrnes, 2013) to estimate the common factor model. The R code is available from the first author upon request.
11.
The only difference compared to the SRMR for a common factor model is that the discrepancies between the observed correlations and the model-implied correlations are zero for pairs of indicators belonging to the same construct.
12.
Since the relative GoF has been shown to be a poorly performing statistic, it is not further discussed here.
13.
Our initial simulation results suggest that such a threshold value will probably lie between 0.12 and 0.14.
14.
See also Dijkstra (2010, in press) for a sketch of Wold’s philosophy when developing PLS.
References
|
Antonakis, J., Bendahan, S., Jacquart, P., Lalive, R. (2010). On making causal claims: A review and recommendations. Leadership Quarterly, 21(6), 1086–1120. doi:10.1016/j.leaqua.2010.10.010 Google Scholar | Crossref | ISI | |
|
Areskoug, B. (1982). The first canonical correlation: Theoretical PLS analysis and simulation experiments. In Jöreskog, K. G., Wold, H. (Eds.), Systems under indirect observation: Causality, structure, prediction (pp. 95-117). Amsterdam, Netherlands: North Holland. Google Scholar | |
|
Atinc, G., Simmering, M. J., Kroll, M. J. (2012). Control variable use and reporting in macro and micro management research. Organizational Research Methods, 15(1), 57–74. doi:10.1177/1094428110397773 Google Scholar | SAGE Journals | ISI | |
|
Barclay, D., Higgins, C., Thompson, R. (1995). The partial least squares (PLS) approach to causal modeling: Personal computer adoption and use as an illustration. Technology Studies, 2(2), 285–309. Google Scholar | |
|
Becker, J.-M., Rai, A., Rigdon, E. E. (2013, December). Predictive validity and formative measurement in structural equation modeling: Embracing practical relevance. Paper presented at the ICIS, Milan, Italy. Google Scholar | |
|
Becker, J.-M., Rai, A., Ringle, C. M., Völckner, F. (2013). Discovering unobserved heterogeneity in structural equation models to avert validity threats. MIS Quarterly, 37(3), 665–694. Google Scholar | Crossref | ISI | |
|
Bentler, P. M. (1976). Multistructure statistical model applied to factor analysis. Multivariate Behavioral Research, 11(1), 3–25. doi:10.1207/s15327906mbr1101_1 Google Scholar | Crossref | Medline | ISI | |
|
Bentler, P. M., Huang, W. (in press). On components, latent variables, PLS and simple methods: Reactions to Ridgon’s rethinking of PLS. Long Range Planning. doi:10.1016/j.lrp.2014.02.005 Google Scholar | |
|
Bollen, K. A. (1989). Structural equations with latent variables. New York, NY: John Wiley. Google Scholar | Crossref | |
|
Bollen, K. A., Long, J. S. (1993). Testing structural equation models. Newbury Park, CA: Sage. Google Scholar | |
|
Byrne, B. M. (1998). Structural equation modeling with LISREL, PRELIS, and SIMPLIS: Basic concepts, applications, and programming. New York, NY: Psychology Press. Google Scholar | |
|
Canty, A., Ripley, B. (2013). boot: Bootstrap R (S-Plus) functions (R package version 1.3-9). Vienna, Austria: R Foundation for Statistical Computing. Google Scholar | |
|
Chin, W. W., Frye, T. A. (2003). PLS graph–version 3.0. Houston, TX: Soft Modeling. Google Scholar | |
|
Chin, W. W., Marcolin, B. L., Newsted, P. R. (2003). A partial least squares latent variable modeling approach for measuring interaction effects: Results from a Monte Carlo simulation study and an electronic-mail emotion/adoption study. Information Systems Research, 14(2), 189–217. doi:10.1287/isre.14.2.189.16018 Google Scholar | Crossref | ISI | |
|
Cliff, N. (1983). Some cautions concerning the application of causal modeling methods. Multivariate Behavioral Research, 18(1), 115–126. doi:10.1207/s15327906mbr1801_7 Google Scholar | Crossref | Medline | ISI | |
|
Creager, J. A., Valentine, L. D. (1962). Regression analysis of linear composite variance. Psychometrika, 27(1), 31–37. doi:10.1007/BF02289662 Google Scholar | Crossref | ISI | |
|
Davison, A. C., Hinkley, D. V. (1997). Bootstrap methods and their applications. Cambridge, UK: Cambridge University Press. Google Scholar | Crossref | |
|
DiCiccio, T. J., Efron, B. (1996). Bootstrap confidence intervals. Statistical Science, 11(3), 189–212. Google Scholar | Crossref | ISI | |
|
Dijkstra, T. K. (1983). Some comments on maximum likelihood and partial least squares methods. Journal of Econometrics, 22 (1/2), 67–90. doi:10.1016/0304-4076(83)90094-5 Google Scholar | Crossref | ISI | |
|
Dijkstra, T. K. (2010). Latent variables and indices: Herman Wold’s basic design and partial least squares. In Esposito Vinzi, V., Chin, W. W., Henseler, J., Wang, H. (Eds.), Handbook of partial least squares: Concepts, methods and applications (Springer handbooks of computational statistics series, vol. 2) (pp. 23–46). Heidelberg, Germany: Springer. doi:10.1007/978-3-540-32827-8_2 Google Scholar | Crossref | |
|
Dijkstra, T. K. (in press). PLS’ Janus face. Long Range Planning. doi:10.1016/j.lrp.2014.02.004 Google Scholar | |
|
Dijkstra, T. K., Henseler, J. (2012). Consistent and asymptotically normal PLS-estimators for linear structural equations (Working paper). Retrieved from http://www.rug.nl/staff/t.k.dijkstra/research. Google Scholar | |
|
Dijkstra, T. K., Schermelleh-Engel, K. (in press). Consistent partial least squares for nonlinear structural equation models. Psychometrika. doi:10.1007/s11336-013-9370-0 Google Scholar | ISI | |
|
Esposito Vinzi, V., Trinchera, L., Amato, S. (2010). PLS path modeling: From foundations to recent developments and open issues for model assessment and improvement. In Esposito Vinzi, V., Chin, W. W., Henseler, J., Wang, H. (Eds.), Handbook of partial least squares: Concepts, methods and applications (Springer handbooks of computational statistics series, vol. 2) (pp. 47–82). Heidelberg, Germany: Springer. doi:10.1007/978-3-540-32827-8_3 Google Scholar | Crossref | |
|
Fornell, C., Bookstein, F. L. (1982). Two structural equation models: LISREL and PLS applied to consumer exit-voice theory. Journal of Marketing Research, 19(4), 440–452. doi:10.2307/3151718 Google Scholar | SAGE Journals | ISI | |
|
Fornell, C., Larcker, D. F. (1981). Evaluating structural equation models with unobservable variables and measurement error. Journal of Marketing Research, 18(1), 39–50. doi:10.2307/3151312 Google Scholar | SAGE Journals | ISI | |
|
Fox, J., Nie, Z., Byrnes, J. (2013). sem: Structural equation models (R package version 3.1-3). Vienna, Austria: R Foundation for Statistical Computing. Google Scholar | |
|
Gefen, D., Rigdon, E. E., Straub, D. W. (2011). Editor’s comment: An update and extension to SEM guidelines for administrative and social science research. MIS Quarterly, 35(2), iii–xiv. Google Scholar | Crossref | ISI | |
|
Gerbing, D. W., Hamilton, J. G. (1994). The surprising viability of a simple alternate estimation procedure for construction of large-scale structural equation measurement models. Structural Equation Modeling: A Multidisciplinary Journal, 1(2), 103–115. doi:10.1080/10705519409539967 Google Scholar | Crossref | ISI | |
|
Goodhue, D. L., Lewis, W., Thompson, R. (2006, January). PLS, small sample size, and statistical power in MIS research. Paper presented at the 39th Annual Hawaii International Conference on System Sciences, Koloa, HI. Google Scholar | Crossref | |
|
Goodhue, D. L., Lewis, W., Thompson, R. (2012). Does PLS have advantages for small sample size or non-normal data? MIS Quarterly, 36(3), 981–1001. Google Scholar | Crossref | ISI | |
|
Greenwald, A. G., Pratkanis, A. R., Leippe, M. R., Baumgardner, M. H. (1986). Under what conditions does theory obstruct research progress? Psychological Review, 93(2), 216–229. doi:10.1037/0033-295X.93.2.216 Google Scholar | Crossref | Medline | ISI | |
|
Hair, J. F., Anderson, R. E., Tatham, R. L., Black, W. C. (1998). Multivariate analysis. Englewood Cliffs, NJ: Prentice Hall. Google Scholar | |
|
Hair, J. F., Hult, G. T. M., Ringle, C. M., Sarstedt, M. (2014). A primer on partial least squares structural equation modeling (PLS-SEM). Thousand Oaks, CA: Sage. Google Scholar | |
|
Hair, J. F., Ringle, C. M., Sarstedt, M. (2011). PLS-SEM: Indeed a silver bullet. Journal of Marketing Theory and Practice, 19(2), 139–151. doi:10.2753/MTP1069-6679190202 Google Scholar | Crossref | |
|
Hair, J. F., Ringle, C. M., Sarstedt, M. (2013). Partial least squares structural equation modeling: Rigorous applications, better results and higher acceptance. Long Range Planning, 46(1-2), 1–12. doi:10.1016/j.lrp.2013.01.001 Google Scholar | Crossref | ISI | |
|
Hair, J. F., Sarstedt, M., Pieper, T. M., Ringle, C. M. (2012). Applications of partial least squares path modeling in management journals: A review of past practices and recommendations for future applications. Long Range Planning, 45(5-6), 320–340. doi:10.1016/j.lrp.2012.09.008 Google Scholar | Crossref | ISI | |
|
Hair, J. F., Sarstedt, M., Ringle, C. M., Mena, J. A. (2012). An assessment of the use of partial least squares structural equation modeling in marketing research. Journal of the Academy of Marketing Science, 40(3), 414–433. doi:10.1007/s11747-011-0261-6 Google Scholar | Crossref | ISI | |
|
Hanafi, M., Qannari, M. E. (2005). An alternative algorithm to the PLS B problem. Computational Statistics & Data Analysis, 48(1), 63–67. doi:10.1016/j.csda.2003.09.007 Google Scholar | Crossref | ISI | |
|
Hayes, A. F., Scharkow, M. (2013). The relative trustworthiness of inferential tests of the indirect effect in statistical mediation analysis: Does method really matter? Psychological Science, 24(10), 1918–1927. doi:10.1177/0956797613480187 Google Scholar | SAGE Journals | ISI | |
|
Heene, M., Hilbert, S., Freudenthaler, H. H., Bühner, M. (2012). Sensitivity of SEM fit indexes with respect to violations of uncorrelated errors. Structural Equation Modeling: A Multidisciplinary Journal, 19(1), 36–50. doi:10.1080/10705511.2012.634710 Google Scholar | Crossref | ISI | |
|
Henseler, J. (2010). On the convergence of the partial least squares path modeling algorithm. Computational Statistics, 25(1), 107–120. doi:10.1007/s00180-009-0164-x Google Scholar | Crossref | ISI | |
|
Henseler, J. (2012). Why generalized structured component analysis is not universally preferable to structural equation modeling. Journal of the Academy of Marketing Science, 40(3), 402–413. doi:10.1007/s11747-011-0298-6 Google Scholar | Crossref | ISI | |
|
Henseler, J., Chin, W. W. (2010). A comparison of approaches for the analysis of interaction effects between latent variables using partial least squares path modeling. Structural Equation Modeling: A Multidisciplinary Journal, 17(1), 82–109. doi:10.1080/10705510903439003 Google Scholar | Crossref | ISI | |
|
Henseler, J., Ringle, C. M., Sinkovics, R. R. (2009). The use of partial least squares path modeling in international marketing. In Sinkovics, R. R., Ghauri, P. N. (Eds.), Advances in international marketing (Vol. 20, pp. 277–320). Bingley, UK: Emerald. doi:10.1108/S1474-7979(2009)0000020014 Google Scholar | Crossref | |
|
Henseler, J., Sarstedt, M. (2013). Goodness-of-fit indices for partial least squares path modeling. Computational Statistics, 28(2), 565–580. doi:10.1007/s00180-012-0317-1 Google Scholar | Crossref | ISI | |
|
Homburg, C., Dobratz, A. (1992). Covariance structure analysis via specification searches. Statistical Papers, 33(1), 119–142. doi:10.1007/BF02925318 Google Scholar | Crossref | |
|
Hu, L.-T., Bentler, P. M. (1998). Fit indices in covariance structure modeling: Sensitivity to underparameterized model misspecification. Psychological Methods, 3(4), 424–453. doi:10.1037/1082-989X.3.4.424 Google Scholar | Crossref | ISI | |
|
Hu, L.-T., Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55. doi:10.1080/10705519909540118 Google Scholar | Crossref | ISI | |
|
Hui, B. S., Wold, H. (1982). Consistency and consistency at large of partial least squares estimates. In Jöreskog, K. G., Wold, H. (Eds.), Systems under indirect observation, part II (pp. 119–130). Amsterdam, Netherlands: North Holland. Google Scholar | |
|
Hulland, J. (1999). Use of partial least squares (PLS) in strategic management research: A review of four recent studies. Strategic Management Journal, 20(2), 195–204. doi:10.1002/(SICI)1097-0266(199902)20:2<195::AID-SMJ13>3.0.CO;2-7 Google Scholar | Crossref | ISI | |
|
Hulland, J., Ryan, M. J., Rayner, R. K. (2010). Modeling customer satisfaction: A comparative performance evaluation of covariance structure analysis versus partial least squares. In Esposito Vinzi, V., Chin, W. W., Henseler, J., Wang, H. (Eds.), Handbook of partial least squares: Concepts, methods and applications (Springer handbooks of computational statistics series, vol. 2) (pp. 307–325). Heidelberg, Germany: Springer. doi:10.1007/978-3-540-32827-8_15 Google Scholar | Crossref | |
|
Hwang, H., Malhotra, N. K., Kim, Y., Tomiuk, M. A., Hong, S. (2010). A comparative study on parameter recovery of three approaches to structural equation modeling. Journal of Marketing Research, 47(4), 699–712. doi:10.1509/jmkr.47.4.699 Google Scholar | Crossref | ISI | |
|
Jagpal, H. S. (1982). Multicollinearity in structural equation models with unobservable variables. Journal of Marketing Research, 19(4), 431–439. doi:10.2307/3151717 Google Scholar | SAGE Journals | ISI | |
|
Jakobowicz, E., Derquenne, C. (2007). A modified PLS path modeling algorithm handling reflective categorical variables and a new model building strategy. Computational Statistics & Data Analysis, 51(8), 3666–3678. doi:10.1016/j.csda.2006.12.004 Google Scholar | Crossref | ISI | |
|
Jöreskog, K. G., Sörbom, D. (1993). Lisrel 8: Structured equation modeling with the Simplis command language. Skokie, IL: Scientific Software International. Google Scholar | |
|
Jöreskog, K. G., Wold, H. (1982). The ML and PLS techniques for modeling with latent variables: historical and comparative aspects. In Wold, H., Jöreskog, K. G. (Eds.), Systems under indirect observation, part I (pp. 263–270). Amsterdam, Netherlands: North Holland. Google Scholar | |
|
Kaplan, A. (1946). Definition and specification of meaning. Journal of Philosophy, 43(11), 281–288. doi:10.2307/2019221 Google Scholar | Crossref | |
|
Ketterlinus, R. D., Bookstein, F. L., Sampson, P. D., Lamb, M. E. (1989). Partial least squares analysis in developmental psychopathology. Development and Psychopathology, 1(2), 351–371. doi:10.1017/S0954579400000523 Google Scholar | Crossref | |
|
Krijnen, W. P., Dijkstra, T. K., Gill, R. D. (1998). Conditions for factor (in)determinacy in factor analysis. Psychometrika, 63(4), 359–367. doi:10.1007/BF02294860 Google Scholar | Crossref | ISI | |
|
Landis, R. S., Beal, D. J., Tesluk, P. E. (2000). A comparison of approaches to forming composite measures in structural equation models. Organizational Research Methods, 3(2), 186–207. doi:10.1177/109442810032003 Google Scholar | SAGE Journals | ISI | |
|
Lee, L., Petter, S., Fayard, D., Robinson, S. (2011). On the use of partial least squares path modeling in accounting research. International Journal of Accounting Information Systems, 12(4), 305–328. doi:10.1016/j.accinf.2011.05.002 Google Scholar | Crossref | |
|
Ley, P. (1972). Quantitative aspects of psychological assessment. London, UK: Duckworth. Google Scholar | |
|
Lohmöller, J.-B. (1988). The PLS program system: Latent variables path analysis with partial least squares estimation. Multivariate Behavioral Research, 23(1), 125–127. doi:10.1207/s15327906mbr2301_7 Google Scholar | Crossref | Medline | ISI | |
|
Lohmöller, J.-B. (1989). Latent variable path modeling with partial least squares. Heidelberg, Germany: Physica. Google Scholar | Crossref | |
|
Lohmöller, J.-B., Wold, H. (1980, June). Three-mode path models with latent variables and partial least squares (PLS) parameter estimation. Paper presented at the European Meeting of the Psychometric Society, Groningen, Netherlands. Google Scholar | |
|
Lu, I. R. R., Kwan, E., Thomas, D. R., Cedzynski, M. (2011). Two new methods for estimating structural equation models: An illustration and a comparison with two established methods. International Journal of Research in Marketing, 28(3), 258–268. doi:10.1016/j.ijresmar.2011.03.006 Google Scholar | Crossref | ISI | |
|
Maraun, M. D., Halpin, P. F. (2008). Manifest and latent variates. Measurement: Interdisciplinary Research and Perspectives, 6(1-2), 113–117. doi:10.1080/15366360802035596 Google Scholar | Crossref | |
|
Marcoulides, G. A., Saunders, C. (2006). PLS: A silver bullet? MIS Quarterly, 30(2), iii–ix. Google Scholar | Crossref | ISI | |
|
McDonald, R. P. (1996). Path analysis with composite variables. Multivariate Behavioral Research, 31(2), 239–270. doi:10.1207/s15327906mbr3102_5 Google Scholar | Crossref | Medline | ISI | |
|
Monecke, A., Leisch, F. (2012). semPLS: Structural equation modeling using partial least squares. Journal of Statistical Software, 48(3), 1–32. Google Scholar | Crossref | ISI | |
|
Mooney, C. Z. (1997). Monte Carlo simulation. Thousand Oaks, CA: Sage. Google Scholar | Crossref | |
|
Mulaik, S. A. (2010). The foundations of factor analysis (2nd ed.). Boca Raton, FL: Chapman & Hall/CRC Press. Google Scholar | |
|
Muthén, L. K., Muthén, B. O. (2006). Mplus version 4 user’s guide. Los Angeles, CA: Muthén & Muthén. Google Scholar | |
|
Paxton, P., Curran, P. J., Bollen, K. A., Kirby, J., Chen, F. (2001). Monte Carlo experiments: Design and implementation. Structural Equation Modeling: A Multidisciplinary Journal, 8(2), 287–312. doi:10.1207/S15328007SEM0802_7 Google Scholar | Crossref | ISI | |
|
Peng, D. X., Lai, F. (2012). Using partial least squares in operations management research: A practical guideline and summary of past research. Journal of Operations Management, 30(6), 467–480. doi:10.1016/j.jom.2012.06.002 Google Scholar | Crossref | ISI | |
|
R Core Team . (2013). R: A language and environment
for statistical computing. Vienna, Austria:
R Foundation for Statistical
Computing. Google Scholar | |
|
Rai, A., Goodhue, D., Henseler, J., Thompson, R. (2013, August). To PLS or not to PLS: That is the question. Paper presented at the 19th Americas Conference on Information Systems, Chicago, IL. Google Scholar | |
|
Rego, L. L. (1998). The relationship market structure-market efficiency from a customer satisfaction perspective. Advances in Consumer Research, 25(1), 132–138. Google Scholar | |
|
Reinartz, W. J., Haenlein, M., Henseler, J. (2009). An empirical comparison of the efficacy of covariance-based and variance-based SEM. International Journal of Research in Marketing, 26(4), 332–344. doi:10.1016/j.ijresmar.2009.08.001 Google Scholar | Crossref | ISI | |
|
Rigdon, E. E. (2012). Rethinking partial least squares path modeling: In praise of simple methods. Long Range Planning, 45(5-6), 341–358. doi:10.1016/j.lrp.2012.09.010 Google Scholar | Crossref | ISI | |
|
Rigdon, E. E. (2013). Partial least squares path modeling. In Hancock, G. R., Mueller, R. O. (Eds.), Structural equation modeling: A second course (2nd ed., pp. 81–116). Greenwich, CT: Information Age. Google Scholar | |
|
Rigdon, E. E. (in press). Rethinking partial least squares path modeling: Breaking chains and forging ahead. Long Range Planning. doi:10.1016/j.lrp.2014.02.003 Google Scholar | ISI | |
|
Ringle, C. M., Sarstedt, M., Schlittgen, R. (2014). Genetic algorithm segmentation in partial least squares structural equation modeling. OR Spectrum, 36(1), 251–276. doi:10.1007/s00291-013-0320-0 Google Scholar | Crossref | ISI | |
|
Ringle, C. M., Sarstedt, M., Schlittgen, R., Taylor, C. R. (2013). PLS path modeling and evolutionary segmentation. Journal of Business Research, 66(9), 1318–1324. doi:10.1016/j.jbusres.2012.02.031 Google Scholar | Crossref | ISI | |
|
Ringle, C. M., Sarstedt, M., Straub, D. W. (2012). A critical look at the use of PLS-SEM in MIS Quarterly. MIS Quarterly, 36(1), iii–xiv. Google Scholar | Crossref | ISI | |
|
Ringle, C. M., Wende, S., Will, A. (2005). SmartPLS 2.0. Retrieved from www.smartpls.de. Google Scholar | |
|
Rönkkö, M., Evermann, J. (2013). A critical examination of common beliefs about partial least squares path modeling. Organizational Research Methods, 16(3), 425–448. doi:10.1177/1094428112474693 Google Scholar | SAGE Journals | ISI | |
|
Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48(2), 1–36. Google Scholar | Crossref | ISI | |
|
Rubinstein, R. Y., Kroese, D. P. (2008). Simulation and the Monte Carlo method (2nd ed.). New York, NY: John Wiley. Google Scholar | |
|
Sarstedt, M., Becker, J.-M., Ringle, C. M., Schwaiger, M. (2011). Uncovering and treating unobserved heterogeneity with FIMIX-PLS: Which model selection criterion provides an appropriate number of segments? Schmalenbach Business Review, 63(1), 34–62. Google Scholar | Crossref | |
|
Sarstedt, M., Ringle, C. M., Henseler, J., Hair, J. F. (in press). On the emancipation of PLS-SEM: A commentary on Rigdon (2012). Long Range Planning. doi:10.1016/j.lrp.2014.02.007 Google Scholar | ISI | |
|
Schönemann, P. H., Wang, M.-M. (1972). Some new results on factor indeterminacy. Psychometrika, 37(1), 61–91. doi:10.1007/BF02291413 Google Scholar | Crossref | ISI | |
|
Sosik, J. J., Kahai, S. S., Piovoso, M. J. (2009). Silver bullet or voodoo statistics? A primer for using the partial least squares data analytic technique in group and organization research. Group Organization Management, 34(1), 5–36. doi:10.1177/1059601108329198 Google Scholar | SAGE Journals | ISI | |
|
Temme, D., Kreis, H., Hildebrandt, L. (2010). A comparison of current PLS path modeling software: Features, ease-of-use, and performance. In Esposito Vinzi, V., Chin, W. W., Henseler, J., Wang, H. (Eds.), Handbook of partial least squares: Concepts, methods and applications (Springer handbooks of computational statistics series, vol. 2) (pp. 737–756). Heidelberg, Germany: Springer. doi:10.1007/978-3-540-32827-8_32 Google Scholar | Crossref | |
|
Tenenhaus, A., Tenenhaus, M. (2011). Regularized generalized canonical correlation analysis. Psychometrika, 76(2), 257–284. doi:10.1007/S11336-011-9206-8 Google Scholar | Crossref | ISI | |
|
Tenenhaus, M. (2008). Component-based structural equation modelling. Total Quality Management & Business Excellence, 19(7), 871–886. doi:10.1080/14783360802159543 Google Scholar | Crossref | ISI | |
|
Tenenhaus, M., Esposito Vinzi, V., Chatelin, Y.-M., Lauro, C. (2005). PLS path modeling. Computational Statistics & Data Analysis, 48(1), 159–205. doi:10.1016/j.csda.2004.03.005 Google Scholar | Crossref | ISI | |
|
Ullman, J. B. (2006). Structural equation modeling: Reviewing the basics and moving forward. Journal of Personality Assessment, 87(1), 35–50. doi:10.1207/s15327752jpa8701_03 Google Scholar | Crossref | |
|
Vilares, M. J., Almeida, M. H., Coelho, P. S. (2010). Comparison of likelihood and PLS estimators for structural equation modeling: A simulation with customer satisfaction data. In Esposito Vinzi, V., Chin, W. W., Henseler, J., Wang, H. (Eds.), Handbook of partial least squares: Concepts, methods and applications (Springer handbooks of computational statistics series, vol. 2) (pp. 289–305). Heidelberg, Germany: Springer. doi:10.1007/978-3-540-32827-8_14 Google Scholar | Crossref | |
|
Vilares, M. J., Coelho, P. S. (2013). Likelihood and PLS estimators for structural equation modeling: An assessment of sample size, skewness and model misspecification effects. In Lita da Silva, J., Caeiro, F., Natário, I., Braumann, C. A. (Eds.), Advances in regression, survival analysis, extreme values, Markov processes and other statistical applications (pp. 11–33). Berlin, Germany: Springer. doi:10.1007/978-3-642-34904-1_2 Google Scholar | Crossref | |
|
Wold, H. (1974). Causal flows with latent variables: Partings of ways in the light of NIPALS modelling. European Economic Review, 5(1), 67–86. doi:10.1016/0014-2921(74)90008-7 Google Scholar | Crossref | ISI | |
|
Wold, H. (1979). Model construction and evaluation when theoretical knowledge is scarce: An example of the use of partial least squares cahiers du département déconométrie. Geneva, Switzerland: Faculté des Sciences Économiques et Sociales Université de Genève. Google Scholar | |
|
Wold, H. (1980). Model construction and evaluation when theoretical knowledge is scarce: Theory and application of PLS. In Kmenta, J., Ramsey, J. B. (Eds.), Evaluation of econometric models (pp. 47–74). New York, NY: Academic Press. Google Scholar | Crossref | |
|
Wold, H. (1982). Soft modeling: The basic design and some extensions. In Jöreskog, K. G., Wold, H. (Eds.), Systems under indirect observations: Part II (pp. 1–54). Amsterdam, Netherlands: North Holland. Google Scholar | |
|
Wold, H. (1985). Partial least squares. In Kotz, S., Johnson, N. L. (Eds.), Encyclopedia of statistical sciences (Vol. 6, pp. 581–591). New York, NY: John Wiley. Google Scholar |
Author Biographies
Jörg Henseler is an associate professor of marketing at the Institute for Management Research, Radboud University Nijmegen, the Netherlands, and a visiting professor at NOVA School of Statistics and Information Management, Universidade Nova de Lisboa, Portugal. His research mainly focuses on variance-based structural equation modeling and customer-firm relationships. E-mail: [email protected]
Theo K. Dijkstra is a professor of econometrics at the Faculty of Economics and Business, University of Groningen, the Netherlands, and an independent researcher. He specializes in multivariate statistics and multi-criteria decision analysis. His PhD thesis on PLS, LISREL and GLS was supervised by Herman Wold. See https://www.researchgate.net/profile/Theo_Dijkstra. E-mail: t.
Marko Sarstedt is a professor of marketing at the Otto-von-Guericke-University Magdeburg (Germany) and an adjunct professor at the University of Newcastle (Australia). His main research interests are the advancement of research methods to further the understanding of consumer behavior. E-mail:marko.
Christian M. Ringle is a professor of management and the director of the Institute of Human Resource Management and Organizations (HRMO) at the Hamburg University of Technology (TUHH), Germany, and he is a visiting professor to the Faculty of Business and Law Professor at the University of Newcastle, Australia. His research mainly addresses strategic management, organizations, marketing, human resource management, and quantitative methods for business and market research. He is co-founder of SmartPLS (http://www.smartpls.com), the leading PLS software. E-mail: [email protected]
Adamantios Diamantopoulos is Chaired Professor of International Marketing at the Department of Business Administration, University of Vienna, Austria. His main research interests are in international marketing and research methodology. E-mail: adamantios.
Detmar W. Straub is a University System of Georgia Regents Professor and the J. Mack Robinson Distinguished Professor of Information Systems at Georgia State University. He has conducted research in the areas of global IT sourcing, computer security, e-Commerce, technological innovation, international IT studies, and IS research methods. He is the immediate past Editor-in-Chief of MIS Quarterly (5 year term from 2008-2012). For lifetime achievement in the field of information systems, he was awarded a LEO by the Association for Information Systems in 2012. E-mail: [email protected]
David J. Ketchen, Jr. serves as Lowder Eminent Scholar and Professor of Management in the Harbert College of Business at Auburn University in Auburn, Alabama. His research interests include entrepreneurship and franchising, methodological issues in organizational research, strategic supply chain management, and the determinants of superior organizational performance. He served as an associate editor for Organizational Research Methods from 2005 to 2007. E-mail: [email protected]
Joseph F. Hair is founder and a senior scholar of the Coles College of Business DBA Program at Kennesaw State University, Georgia. His research mainly focuses on multivariate analysis methods and their application in business research. He is a Distinguished Fellow of the Academy of Marketing Science and the Society for Marketing Advances. E-mail: [email protected]
G. Tomas M. Hult is professor, Byington Endowed Chair of International Business, and Director of the International Business Center in the Eli Broad College of Business at Michigan State University. His research interests are international business, supply chain management, and marketing strategy. He is Executive Director of the Academy of International Business and an elected Fellow of AIB. E-mail: [email protected]
Roger J. Calantone is the Eli Broad Chaired University Professor of Business at the Eli Broad Graduate School of Management at Michigan State University. He is also adjunct professor of management and economics, and has been recognized as an MSU University Distinguished Faculty. His publications and research are mostly in the areas of product design and development processes, decision support tools for NPD, and organization process metrics. E-mail: [email protected]








