Abstract
New scenarios in data analysis call for new methods capable of fully exploiting the potential of the available data. In the specific context of massive data, this means tackling two principal issues: the computational cost of applying a method, since convergence must be fast, and a level of accuracy that ensures precise estimation. We explore the applicability of the partial least squares structural equation modelling (PLS-SEM) algorithm to a massive data context. Considering the classical bootstrap procedure used to validate model coefficients, we show that bootstrapping becomes very expensive computationally once a sample size becomes massive. We consequently adapted the subsampled double bootstrap (SDB) algorithm to the PLS-SEM context to reduce the computational cost without sacrificing accuracy in confidence interval estimates, using a straightforward procedure that is easy to implement. The accuracy and the speed of convergence of the SDB PLS-SEM are demonstrated in simulation studies, and the new method is successfully tested with a model of European internet use.
| Original language | English |
|---|---|
| Number of pages | 37 |
| Journal | Quality and Quantity |
| DOIs | |
| Publication status | Published - 24 Nov 2025 |
Keywords
- PLS-SEM
- Big Data
- bootstrap procedure
- inference accuracy
- computational cost
- massive secondary data
- internet use
Fingerprint
Dive into the research topics of 'Bootstrapping partial least squares structural equation modelling with massive data'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver